Commit Graph

45 Commits

Author SHA1 Message Date
Jeff Moyer ed486f925a io_uring: Fix probe of disabled operations
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit 3e05b222382ec67dce7358d50b6006e91d028d8b
Author: Gabriel Krisman Bertazi <krisman@suse.de>
Date:   Tue Jun 18 22:06:18 2024 -0400

    io_uring: Fix probe of disabled operations
    
    io_probe checks io_issue_def->not_supported, but we never really set
    that field, as we mark non-supported functions through a specific ->prep
    handler.  This means we end up returning IO_URING_OP_SUPPORTED, even for
    disabled operations.  Fix it by just checking the prep handler itself.
    
    Fixes: 66f4af93da ("io_uring: add support for probing opcodes")
    Signed-off-by: Gabriel Krisman Bertazi <krisman@suse.de>
    Link: https://lore.kernel.org/r/20240619020620.5301-2-krisman@suse.de
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-12-02 11:21:29 -05:00
Jeff Moyer 5e1fd9dc9c io_uring/rw: Free iovec before cleaning async data
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit e112311615a24e1618a591c73506571dc304eb8d
Author: Breno Leitao <leitao@debian.org>
Date:   Thu May 30 07:23:39 2024 -0700

    io_uring/rw: Free iovec before cleaning async data
    
    kmemleak shows that there is a memory leak in io_uring read operation,
    where a buffer is allocated at iovec import, but never de-allocated.
    
    The memory is allocated at io_async_rw->free_iovec, but, then
    io_async_rw is kfreed, taking the allocated memory with it. I saw this
    happening when the read operation fails with -11 (EAGAIN).
    
    This is the kmemleak splat.
    
        unreferenced object 0xffff8881da591c00 (size 256):
    ...
          backtrace (crc 7a15bdee):
            [<00000000256f2de4>] __kmalloc+0x2d6/0x410
            [<000000007a9f5fc7>] iovec_from_user.part.0+0xc6/0x160
            [<00000000cecdf83a>] __import_iovec+0x50/0x220
            [<00000000d1d586a2>] __io_import_iovec+0x13d/0x220
            [<0000000054ee9bd2>] io_prep_rw+0x186/0x340
            [<00000000a9c0372d>] io_prep_rwv+0x31/0x120
            [<000000001d1170b9>] io_prep_readv+0xe/0x30
            [<0000000070b8eb67>] io_submit_sqes+0x1bd/0x780
            [<00000000812496d4>] __do_sys_io_uring_enter+0x3ed/0x5b0
            [<0000000081499602>] do_syscall_64+0x5d/0x170
            [<00000000de1c5a4d>] entry_SYSCALL_64_after_hwframe+0x76/0x7e
    
    This occurs because the async data cleanup functions are not set for
    read/write operations. As a result, the potentially allocated iovec in
    the rw async data is not freed before the async data is released,
    leading to a memory leak.
    
    With this following patch, kmemleak does not show the leaked memory
    anymore, and all liburing tests pass.
    
    Fixes: a9165b83c193 ("io_uring/rw: always setup io_async_rw for read/write requests")
    Signed-off-by: Breno Leitao <leitao@debian.org>
    Link: https://lore.kernel.org/r/20240530142340.1248216-1-leitao@debian.org
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-12-02 11:12:47 -05:00
Jeff Moyer 3309b0778b io_uring/net: add provided buffer support for IORING_OP_SEND
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit ac5f71a3d9d7eb540f6bf7e794eb4a3e4c3f11dd
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Feb 19 10:46:44 2024 -0700

    io_uring/net: add provided buffer support for IORING_OP_SEND
    
    It's pretty trivial to wire up provided buffer support for the send
    side, just like how it's done the receive side. This enables setting up
    a buffer ring that an application can use to push pending sends to,
    and then have a send pick a buffer from that ring.
    
    One of the challenges with async IO and networking sends is that you
    can get into reordering conditions if you have more than one inflight
    at the same time. Consider the following scenario where everything is
    fine:
    
    1) App queues sendA for socket1
    2) App queues sendB for socket1
    3) App does io_uring_submit()
    4) sendA is issued, completes successfully, posts CQE
    5) sendB is issued, completes successfully, posts CQE
    
    All is fine. Requests are always issued in-order, and both complete
    inline as most sends do.
    
    However, if we're flooding socket1 with sends, the following could
    also result from the same sequence:
    
    1) App queues sendA for socket1
    2) App queues sendB for socket1
    3) App does io_uring_submit()
    4) sendA is issued, socket1 is full, poll is armed for retry
    5) Space frees up in socket1, this triggers sendA retry via task_work
    6) sendB is issued, completes successfully, posts CQE
    7) sendA is retried, completes successfully, posts CQE
    
    Now we've sent sendB before sendA, which can make things unhappy. If
    both sendA and sendB had been using provided buffers, then it would look
    as follows instead:
    
    1) App queues dataA for sendA, queues sendA for socket1
    2) App queues dataB for sendB queues sendB for socket1
    3) App does io_uring_submit()
    4) sendA is issued, socket1 is full, poll is armed for retry
    5) Space frees up in socket1, this triggers sendA retry via task_work
    6) sendB is issued, picks first buffer (dataA), completes successfully,
       posts CQE (which says "I sent dataA")
    7) sendA is retried, picks first buffer (dataB), completes successfully,
       posts CQE (which says "I sent dataB")
    
    Now we've sent the data in order, and everybody is happy.
    
    It's worth noting that this also opens the door for supporting multishot
    sends, as provided buffers would be a prerequisite for that. Those can
    trigger either when new buffers are added to the outgoing ring, or (if
    stalled due to lack of space) when space frees up in the socket.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 17:34:44 -05:00
Jeff Moyer 5541c9a578 io_uring: drop ->prep_async()
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit e10677a8f6980dbae2e866b8320d90bae07e87ee
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 20:48:38 2024 -0600

    io_uring: drop ->prep_async()
    
    It's now unused, drop the code related to it. This includes the
    io_issue_defs->manual alloc field.
    
    While in there, and since ->async_size is now being used a bit more
    frequently and in the issue path, move it to io_issue_defs[].
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:55:44 -05:00
Jeff Moyer c22e5fd4b4 io_uring/uring_cmd: switch to always allocating async data
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit d10f19dff56eac5ae44dc270336b18071a8bd51c
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 20:41:58 2024 -0600

    io_uring/uring_cmd: switch to always allocating async data
    
    Basic conversion ensuring async_data is allocated off the prep path. Adds
    a basic alloc cache as well, as passthrough IO can be quite high in rate.
    
    Tested-by: Anuj Gupta <anuj20.g@samsung.com>
    Reviewed-by: Anuj Gupta <anuj20.g@samsung.com>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:53:44 -05:00
Jeff Moyer 8842cece02 io_uring/net: move connect to always using async data
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit e2ea5a7069133c01fe3dbda95d77af7f193a1a52
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 20:37:22 2024 -0600

    io_uring/net: move connect to always using async data
    
    While doing that, get rid of io_async_connect and just use the generic
    io_async_msghdr. Both of them have a struct sockaddr_storage in there,
    and while io_async_msghdr is bigger, if the same type can be used then
    the netmsg_cache can get reused for connect as well.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:52:44 -05:00
Jeff Moyer dbe77cf6f1 io_uring/rw: always setup io_async_rw for read/write requests
JIRA: https://issues.redhat.com/browse/RHEL-64867
Conflicts: Convert READ/WRITE -> ITER_DEST/ITER_SOURCE.

commit a9165b83c1937eeed1f0c731468216d6371d647f
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 16:13:01 2024 -0600

    io_uring/rw: always setup io_async_rw for read/write requests
    
    read/write requests try to put everything on the stack, and then alloc
    and copy if a retry is needed. This necessitates a bunch of nasty code
    that deals with intermediate state.
    
    Get rid of this, and have the prep side setup everything that is needed
    upfront, which greatly simplifies the opcode handlers.
    
    This includes adding an alloc cache for io_async_rw, to make it cheap
    to handle.
    
    In terms of cost, this should be basically free and transparent. For
    the worst case of {READ,WRITE}_FIXED which didn't need it before,
    performance is unaffected in the normal peak workload that is being
    used to test that. Still runs at 122M IOPS.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:48:44 -05:00
Jeff Moyer 34fd898903 io_uring/net: get rid of ->prep_async() for send side
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit 50220d6ac8ff31eb065fba818e960f549fb89d4d
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 08:09:47 2024 -0600

    io_uring/net: get rid of ->prep_async() for send side
    
    Move the io_async_msghdr out of the issue path and into prep handling,
    e it's now done unconditionally and hence does not need to be part
    of the issue path. This means any usage of io_sendrecv_prep_async() and
    io_sendmsg_prep_async(), and hence the forced async setup path is now
    unified with the normal prep setup.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:43:44 -05:00
Jeff Moyer 8ffe89e94d io_uring/net: get rid of ->prep_async() for receive side
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit c6f32c7d9e09bf1368447e9a29e869193ecbb756
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Mar 18 07:36:03 2024 -0600

    io_uring/net: get rid of ->prep_async() for receive side
    
    Move the io_async_msghdr out of the issue path and into prep handling,
    since it's now done unconditionally and hence does not need to be part
    of the issue path. This reduces the footprint of the multishot fast
    path of multiple invocations of ->issue() per prep, and also means that
    using ->prep_async() can be dropped for recvmsg asthis is now done via
    setup on the prep side.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:42:44 -05:00
Jeff Moyer 8620a3c4b1 io_uring/net: switch io_recv() to using io_async_msghdr
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit 4a3223f7bfda14c532856152b12aace525cf8079
Author: Jens Axboe <axboe@kernel.dk>
Date:   Tue Mar 5 15:39:16 2024 -0700

    io_uring/net: switch io_recv() to using io_async_msghdr
    
    No functional changes in this patch, just in preparation for carrying
    more state than what is available now, if necessary.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:38:44 -05:00
Jeff Moyer ef97076384 io_uring/net: switch io_send() and io_send_zc() to using io_async_msghdr
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit 54cdcca05abde32acc3233950ddc79d8be25515f
Author: Jens Axboe <axboe@kernel.dk>
Date:   Tue Mar 5 09:34:21 2024 -0700

    io_uring/net: switch io_send() and io_send_zc() to using io_async_msghdr
    
    No functional changes in this patch, just in preparation for carrying
    more state then what is being done now, if necessary. While unifying
    some of this code, add a generic send setup prep handler that they can
    both use.
    
    This gets rid of some manual msghdr and sockaddr on the stack, and makes
    it look a bit more like the sendmsg/recvmsg variants. Going forward, more
    can get unified on top.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 16:37:44 -05:00
Jeff Moyer 477d82a904 io_uring: add support for ftruncate
JIRA: https://issues.redhat.com/browse/RHEL-64867

commit b4bb1900c12e6a0fe11ff51e1aa6eea19a4ad635
Author: Tony Solomonik <tony.solomonik@gmail.com>
Date:   Fri Feb 2 14:17:24 2024 +0200

    io_uring: add support for ftruncate
    
    Adds support for doing truncate through io_uring, eliminating
    the need for applications to roll their own thread pool or offload
    mechanism to be able to do non-blocking truncates.
    
    Signed-off-by: Tony Solomonik <tony.solomonik@gmail.com>
    Link: https://lore.kernel.org/r/20240202121724.17461-3-tony.solomonik@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-11-28 15:35:44 -05:00
Jeff Moyer fb1b2f09ee io_uring: enable audit and restrict cred override for IORING_OP_FIXED_FD_INSTALL
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit 16bae3e1377846734ec6b87eee459c0f3551692c
Author: Paul Moore <paul@paul-moore.com>
Date:   Tue Jan 23 16:55:02 2024 -0500

    io_uring: enable audit and restrict cred override for IORING_OP_FIXED_FD_INSTALL
    
    We need to correct some aspects of the IORING_OP_FIXED_FD_INSTALL
    command to take into account the security implications of making an
    io_uring-private file descriptor generally accessible to a userspace
    task.
    
    The first change in this patch is to enable auditing of the FD_INSTALL
    operation as installing a file descriptor into a task's file descriptor
    table is a security relevant operation and something that admins/users
    may want to audit.
    
    The second change is to disable the io_uring credential override
    functionality, also known as io_uring "personalities", in the
    FD_INSTALL command.  The credential override in FD_INSTALL is
    particularly problematic as it affects the credentials used in the
    security_file_receive() LSM hook.  If a task were to request a
    credential override via REQ_F_CREDS on a FD_INSTALL operation, the LSM
    would incorrectly check to see if the overridden credentials of the
    io_uring were able to "receive" the file as opposed to the task's
    credentials.  After discussions upstream, it's difficult to imagine a
    use case where we would want to allow a credential override on a
    FD_INSTALL operation so we are simply going to block REQ_F_CREDS on
    IORING_OP_FIXED_FD_INSTALL operations.
    
    Fixes: dc18b89ab113 ("io_uring/openclose: add support for IORING_OP_FIXED_FD_INSTALL")
    Signed-off-by: Paul Moore <paul@paul-moore.com>
    Link: https://lore.kernel.org/r/20240123215501.289566-2-paul@paul-moore.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 14:33:37 -04:00
Jeff Moyer 179f7a7dba io_uring/openclose: add support for IORING_OP_FIXED_FD_INSTALL
JIRA: https://issues.redhat.com/browse/RHEL-27755
Conflicts: RHEL is missing commit 4e94ddfe2aab ("file: remove
  __receive_fd()"), which changes the calling convention of receive_fd.

commit dc18b89ab113e9c6c7a529316ddf7029fb55132d
Author: Jens Axboe <axboe@kernel.dk>
Date:   Thu Dec 7 20:06:02 2023 -0700

    io_uring/openclose: add support for IORING_OP_FIXED_FD_INSTALL
    
    io_uring can currently open/close regular files or fixed/direct
    descriptors. Or you can instantiate a fixed descriptor from a regular
    one, and then close the regular descriptor. But you currently can't turn
    a purely fixed/direct descriptor into a regular file descriptor.
    
    IORING_OP_FIXED_FD_INSTALL adds support for installing a direct
    descriptor into the normal file table, just like receiving a file
    descriptor or opening a new file would do. This is all nicely abstracted
    into receive_fd(), and hence adding support for this is truly trivial.
    
    Since direct descriptors are only usable within io_uring itself, it can
    be useful to turn them into real file descriptors if they ever need to
    be accessed via normal syscalls. This can either be a transitory thing,
    or just a permanent transition for a given direct descriptor.
    
    By default, new fds are installed with O_CLOEXEC set. The application
    can disable O_CLOEXEC by setting IORING_FIXED_FD_NO_CLOEXEC in the
    sqe->install_fd_flags member.
    
    Suggested-by: Christian Brauner <brauner@kernel.org>
    Reviewed-by: Christian Brauner <brauner@kernel.org>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 14:33:35 -04:00
Jeff Moyer 2aad660a9f io_uring/rw: add separate prep handler for fixed read/write
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit f688944cfb810986c626cb13d95bc666e5c8a36c
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Nov 6 07:43:16 2023 -0700

    io_uring/rw: add separate prep handler for fixed read/write
    
    Rather than sprinkle opcode checks in the generic read/write prep handler,
    have a separate prep handler for the vectored readv/writev operation.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 10:07:34 -04:00
Jeff Moyer afd4a8a0dd io_uring/rw: add separate prep handler for readv/writev
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit 0e984ec88da9747549227900e5215c5e6a1b65ae
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Nov 6 07:41:17 2023 -0700

    io_uring/rw: add separate prep handler for readv/writev
    
    Rather than sprinkle opcode checks in the generic read/write prep handler,
    have a separate prep handler for the vectored readv/writev operation.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 10:06:34 -04:00
Jeff Moyer a2b237b6f9 io_uring: add support for vectored futex waits
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit 8f350194d5cfd7016d4cd44e433df0faa4d4a703
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Jun 12 19:04:32 2023 -0600

    io_uring: add support for vectored futex waits
    
    This adds support for IORING_OP_FUTEX_WAITV, which allows registering a
    notification for a number of futexes at once. If one of the futexes are
    woken, then the request will complete with the index of the futex that got
    woken as the result. This is identical to what the normal vectored futex
    waitv operation does.
    
    Use like IORING_OP_FUTEX_WAIT, except sqe->addr must now contain a
    pointer to a struct futex_waitv array, and sqe->off must now contain the
    number of elements in that array. As flags are passed in the futex_vector
    array, and likewise for the value and futex address(es), sqe->addr2
    and sqe->addr3 are also reserved for IORING_OP_FUTEX_WAITV.
    
    For cancelations, FUTEX_WAITV does not rely on the futex_unqueue()
    return value as we're dealing with multiple futexes. Instead, a separate
    per io_uring request atomic is used to claim ownership of the request.
    
    Waiting on N futexes could be done with IORING_OP_FUTEX_WAIT as well,
    but that punts a lot of the work to the application:
    
    1) Application would need to submit N IORING_OP_FUTEX_WAIT requests,
       rather than just a single IORING_OP_FUTEX_WAITV.
    
    2) When one futex is woken, application would need to cancel the
       remaining N-1 requests that didn't trigger.
    
    While this is of course doable, having a single vectored futex wait
    makes for much simpler application code.
    
    Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 09:51:34 -04:00
Jeff Moyer f33d7fd070 io_uring: add support for futex wake and wait
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit 194bb58c6090e39bd7d9b9c888a079213628e1f6
Author: Jens Axboe <axboe@kernel.dk>
Date:   Thu Jun 8 11:57:40 2023 -0600

    io_uring: add support for futex wake and wait
    
    Add support for FUTEX_WAKE/WAIT primitives.
    
    IORING_OP_FUTEX_WAKE is mix of FUTEX_WAKE and FUTEX_WAKE_BITSET, as
    it does support passing in a bitset.
    
    Similary, IORING_OP_FUTEX_WAIT is a mix of FUTEX_WAIT and
    FUTEX_WAIT_BITSET.
    
    For both of them, they are using the futex2 interface.
    
    FUTEX_WAKE is straight forward, as those can always be done directly from
    the io_uring submission without needing async handling. For FUTEX_WAIT,
    things are a bit more complicated. If the futex isn't ready, then we
    rely on a callback via futex_queue->wake() when someone wakes up the
    futex. From that calback, we queue up task_work with the original task,
    which will post a CQE and wake it, if necessary.
    
    Cancelations are supported, both from the application point-of-view,
    but also to be able to cancel pending waits if the ring exits before
    all events have occurred. The return value of futex_unqueue() is used
    to gate who wins the potential race between cancelation and futex
    wakeups. Whomever gets a 'ret == 1' return from that claims ownership
    of the io_uring futex request.
    
    This is just the barebones wait/wake support. PI or REQUEUE support is
    not added at this point, unclear if we might look into that later.
    
    Likewise, explicit timeouts are not supported either. It is expected
    that users that need timeouts would do so via the usual io_uring
    mechanism to do that using linked timeouts.
    
    The SQE format is as follows:
    
    `addr`          Address of futex
    `fd`            futex2(2) FUTEX2_* flags
    `futex_flags`   io_uring specific command flags. None valid now.
    `addr2`         Value of futex
    `addr3`         Mask to wake/wait
    
    Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 09:50:34 -04:00
Jeff Moyer 05243a1f83 io_uring: add IORING_OP_WAITID support
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit f31ecf671ddc498f20219453395794ff2383e06b
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Jul 10 16:14:37 2023 -0600

    io_uring: add IORING_OP_WAITID support
    
    This adds support for an async version of waitid(2), in a fully async
    version. If an event isn't immediately available, wait for a callback
    to trigger a retry.
    
    The format of the sqe is as follows:
    
    sqe->len                The 'which', the idtype being queried/waited for.
    sqe->fd                 The 'pid' (or id) being waited for.
    sqe->file_index         The 'options' being set.
    sqe->addr2              A pointer to siginfo_t, if any, being filled in.
    
    buf_index, add3, and waitid_flags are reserved/unused for now.
    waitid_flags will be used for options for this request type. One
    interesting use case may be to add multi-shot support, so that the
    request stays armed and posts a notification every time a monitored
    process state change occurs.
    
    Note that this does not support rusage, on Arnd's recommendation.
    
    See the waitid(2) man page for details on the arguments.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 09:47:34 -04:00
Jeff Moyer fee01a0d73 io_uring/rw: add support for IORING_OP_READ_MULTISHOT
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit fc68fcda049108478ee4704d8a3ad3e05cc72fd0
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Sep 11 13:35:42 2023 -0600

    io_uring/rw: add support for IORING_OP_READ_MULTISHOT
    
    This behaves like IORING_OP_READ, except:
    
    1) It only supports pollable files (eg pipes, sockets, etc). Note that
       for sockets, you probably want to use recv/recvmsg with multishot
       instead.
    
    2) It supports multishot mode, meaning it will repeatedly trigger a
       read and fill a buffer when data is available. This allows similar
       use to recv/recvmsg but on non-sockets, where a single request will
       repeatedly post a CQE whenever data is read from it.
    
    3) Because of #2, it must be used with provided buffers. This is
       uniformly true across any request type that supports multishot and
       transfers data, with the reason being that it's obviously not
       possible to pass in a single buffer for the data, as multiple reads
       may very well trigger before an application has a chance to process
       previous CQEs and the data passed from them.
    
    Reviewed-by: Gabriel Krisman Bertazi <krisman@suse.de>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 09:46:34 -04:00
Jeff Moyer 11796f3222 io_uring/rw: mark readv/writev as vectored in the opcode definition
JIRA: https://issues.redhat.com/browse/RHEL-27755

commit d2d778fbf9964e4e5b8d7420eba8ec5ce938e794
Author: Jens Axboe <axboe@kernel.dk>
Date:   Mon Sep 11 13:46:07 2023 -0600

    io_uring/rw: mark readv/writev as vectored in the opcode definition
    
    This is cleaner than gating on the opcode type, particularly as more
    read/write type opcodes may be added.
    
    Then we can use that for the data import, and for __io_read() on
    whether or not we need to copy state.
    
    Reviewed-by: Gabriel Krisman Bertazi <krisman@suse.de>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2024-07-02 09:41:34 -04:00
Jeff Moyer 0454a9fd14 io_uring: Pass whole sqe to commands
JIRA: https://issues.redhat.com/browse/RHEL-12076
Conflicts: RHEL does not yet support ublk, so this backport removes
  those changes from the patch.

commit fd9b8547bc5c34186dc42ea05fb4380d21695374
Author: Breno Leitao <leitao@debian.org>
Date:   Thu May 4 05:18:55 2023 -0700

    io_uring: Pass whole sqe to commands
    
    Currently uring CMD operation relies on having large SQEs, but future
    operations might want to use normal SQE.
    
    The io_uring_cmd currently only saves the payload (cmd) part of the SQE,
    but, for commands that use normal SQE size, it might be necessary to
    access the initial SQE fields outside of the payload/cmd block.  So,
    saves the whole SQE other than just the pdu.
    
    This changes slightly how the io_uring_cmd works, since the cmd
    structures and callbacks are not opaque to io_uring anymore. I.e, the
    callbacks can look at the SQE entries, not only, in the cmd structure.
    
    The main advantage is that we don't need to create custom structures for
    simple commands.
    
    Creates io_uring_sqe_cmd() that returns the cmd private data as a null
    pointer and avoids casting in the callee side.
    Also, make most of ublk_drv's sqe->cmd priv structure into const, and use
    io_uring_sqe_cmd() to get the private structure, removing the unwanted
    cast. (There is one case where the cast is still needed since the
    header->{len,addr} is updated in the private structure)
    
    Suggested-by: Pavel Begunkov <asml.silence@gmail.com>
    Signed-off-by: Breno Leitao <leitao@debian.org>
    Reviewed-by: Keith Busch <kbusch@kernel.org>
    Reviewed-by: Christoph Hellwig <hch@lst.de>
    Reviewed-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/20230504121856.904491-3-leitao@debian.org
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-11-02 15:31:34 -04:00
Jeff Moyer 0ed86416ea io_uring: Split io_issue_def struct
JIRA: https://issues.redhat.com/browse/RHEL-12076
Conflicts: Minor conflicts due to having applied patches out of order.

commit f30bd4d03824fb437bf080c2b2f926cfee3f09d0
Author: Breno Leitao <leitao@debian.org>
Date:   Thu Jan 12 06:44:11 2023 -0800

    io_uring: Split io_issue_def struct
    
    This patch removes some "cold" fields from `struct io_issue_def`.
    
    The plan is to keep only highly used fields into `struct io_issue_def`, so,
    it may be hot in the cache. The hot fields are basically all the bitfields
    and the callback functions for .issue and .prep.
    
    The other less frequently used fields are now located in a secondary and
    cold struct, called `io_cold_def`.
    
    This is the size for the structs:
    
    Before: io_issue_def = 56 bytes
    After: io_issue_def = 24 bytes; io_cold_def = 40 bytes
    
    Signed-off-by: Breno Leitao <leitao@debian.org>
    Reviewed-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/20230112144411.2624698-2-leitao@debian.org
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-11-02 15:31:09 -04:00
Jeff Moyer 0c3b521dde io_uring: Rename struct io_op_def
JIRA: https://issues.redhat.com/browse/RHEL-12076

commit a7dd27828b00be8c0c7520c53baf0b360f4d8bea
Author: Breno Leitao <leitao@debian.org>
Date:   Thu Jan 12 06:44:10 2023 -0800

    io_uring: Rename struct io_op_def
    
    The current io_op_def struct is becoming huge and the name is a bit
    generic.
    
    The goal of this patch is to rename this struct to `io_issue_def`. This
    struct will contain the hot functions associated with the issue code
    path.
    
    For now, this patch only renames the structure, and an upcoming patch
    will break up the structure in two, moving the non-issue fields to a
    secondary struct.
    
    Signed-off-by: Breno Leitao <leitao@debian.org>
    Reviewed-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/20230112144411.2624698-1-leitao@debian.org
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-11-02 15:31:09 -04:00
Jeff Moyer 7219eed4af io_uring,audit: don't log IORING_OP_MADVISE
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit fbe870a72fd1ddc5e08c23764e23e5766f54aa87
Author: Richard Guy Briggs <rgb@redhat.com>
Date:   Wed Feb 1 15:33:33 2023 -0500

    io_uring,audit: don't log IORING_OP_MADVISE
    
    fadvise and madvise both provide hints for caching or access pattern for
    file and memory respectively.  Skip them.
    
    Fixes: 5bd2182d58e9 ("audit,io_uring,io-wq: add some basic audit support to io_uring")
    Signed-off-by: Richard Guy Briggs <rgb@redhat.com>
    Link: https://lore.kernel.org/r/b5dfdcd541115c86dbc774aa9dd502c964849c5f.1675282642.git.rgb@redhat.com
    Acked-by: Paul Moore <paul@paul-moore.com>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:26:14 -04:00
Jeff Moyer de845c40fc io_uring: get rid of double locking
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 11373026f2960390d5e330df4e92735c4265c440
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Dec 7 03:53:34 2022 +0000

    io_uring: get rid of double locking
    
    We don't need to take both uring_locks at once, msg_ring can be split in
    two parts, first getting a file from the filetable of the first ring and
    then installing it into the second one.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/a80ecc2bc99c3b3f2cf20015d618b7c51419a797.1670384893.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:25:21 -04:00
Jeff Moyer ac9e612458 io_uring: dont remove file from msg_ring reqs
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit ef0ec1ad03119b8b46b035dad42bca7d6da7c2e5
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Dec 7 03:53:26 2022 +0000

    io_uring: dont remove file from msg_ring reqs
    
    We should not be messing with req->file outside of core paths. Clearing
    it makes msg_ring non reentrant, i.e. luckily io_msg_send_fd() fails the
    request on failed io_double_lock_ctx() but clearly was originally
    intended to do retries instead.
    
    Cc: stable@vger.kernel.org
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/e5ac9edadb574fe33f6d727cb8f14ce68262a684.1670384893.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:25:18 -04:00
Jeff Moyer 822aa1ceaa io_uring/opdef: remove 'audit_skip' from SENDMSG_ZC
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 11528491c65a493050c682786c6b7cfd9e9b4a8f
Author: Jens Axboe <axboe@kernel.dk>
Date:   Fri Oct 7 12:26:02 2022 -0600

    io_uring/opdef: remove 'audit_skip' from SENDMSG_ZC
    
    The msg variants of sending aren't audited separately, so we should not
    be setting audit_skip for the zerocopy sendmsg variant either.
    
    Fixes: 493108d95f14 ("io_uring/net: zerocopy sendmsg")
    Reported-by: Paul Moore <paul@paul-moore.com>
    Reviewed-by: Paul Moore <paul@paul-moore.com>
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:24:08 -04:00
Jeff Moyer 69ed0cb748 io_uring/net: zerocopy sendmsg
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 493108d95f1464ccd101d4e5cfa7e93f1fc64d47
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:54 2022 +0100

    io_uring/net: zerocopy sendmsg
    
    Add a zerocopy version of sendmsg.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/6aabc4bdfc0ec78df6ec9328137e394af9d4e7ef.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:57 -04:00
Jeff Moyer 7d04818099 io_uring/net: combine fail handlers
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit c4c0009e0b56ef9920020bcade1e45be52653bae
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:53 2022 +0100

    io_uring/net: combine fail handlers
    
    Merge io_send_zc_fail() into io_sendrecv_fail(), saves a few lines of
    code and some headache for following patch.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/e0eba1d577413aef5602cd45f588b9230207082d.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:57 -04:00
Jeff Moyer b5c1eccc26 io_uring/net: rename io_sendzc()
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit b0e9b5517eb12fa80c72e205fe28534c2e2f39b9
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:52 2022 +0100

    io_uring/net: rename io_sendzc()
    
    Simple renaming of io_sendzc*() functions in preparatio to adding
    a zerocopy sendmsg variant.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/265af46829e6076dd220011b1858dc3151969226.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:57 -04:00
Jeff Moyer 4bdf424e42 io_uring/net: support non-zerocopy sendto
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 516e82f0e043a1a0e8d00800ed0ffe2137cf0e7e
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:51 2022 +0100

    io_uring/net: support non-zerocopy sendto
    
    We have normal sends, but what is missing is sendto-like requests. Add
    sendto() capabilities to IORING_OP_SEND by passing in addr just as we do
    for IORING_OP_SEND_ZC.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/69fbd8b2cb830e57d1bf9ec351e9bf95c5b77e3f.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:56 -04:00
Jeff Moyer fd543b3e34 io_uring/net: don't lose partial send_zc on fail
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 5693bcce892d7b8b15a7a92b011d3d40a023b53c
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:49 2022 +0100

    io_uring/net: don't lose partial send_zc on fail
    
    Partial zc send may end up in io_req_complete_failed(), which not only
    would return invalid result but also mask out the notification leading
    to lifetime issues.
    
    Cc: stable@vger.kernel.org
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/5673285b5e83e6ceca323727b4ddaa584b5cc91e.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:56 -04:00
Jeff Moyer dd4be430aa io_uring/net: don't lose partial send/recv on fail
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 7e6b638ed501cced4e472298d6b08dd16346f3a6
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:48 2022 +0100

    io_uring/net: don't lose partial send/recv on fail
    
    Just as with rw, partial send/recv may end up in
    io_req_complete_failed() and loose the result, make sure we return the
    number of bytes processed.
    
    Cc: stable@vger.kernel.org
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/a4ff95897b5419356fca9ea55db91ac15b2975f9.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:55 -04:00
Jeff Moyer a59a82b1fe io_uring/rw: don't lose partial IO result on fail
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 47b4c68660752facfa6247b1fc9ca9d722b8b601
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Sep 21 12:17:47 2022 +0100

    io_uring/rw: don't lose partial IO result on fail
    
    A partially done read/write may end up in io_req_complete_failed() and
    loose the result, make sure we return the number of bytes processed.
    
    Cc: stable@vger.kernel.org
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/05e0879c226bcd53b441bf92868eadd4bf04e2fc.1663668091.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:55 -04:00
Jeff Moyer 4d90be4396 io_uring/net: use async caches for async prep
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 858c293e5d3b7fd3037883fcc0379594517c926c
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Thu Sep 8 13:20:30 2022 +0100

    io_uring/net: use async caches for async prep
    
    send/recv have async_data caches but there are only used from within
    issue handlers. Extend their use also to ->prep_async, should be handy
    with links and IOSQE_ASYNC.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/b9a2264b807582a97ed606c5bfcdc2399384e8a5.1662639236.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:51 -04:00
Jeff Moyer e92ce43ad3 io_uring: add iopoll infrastructure for io_uring_cmd
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 5756a3a7e713bcab705a5f0c810a2b1f7f4ecfaa
Author: Kanchan Joshi <joshi.k@samsung.com>
Date:   Tue Aug 23 21:44:41 2022 +0530

    io_uring: add iopoll infrastructure for io_uring_cmd
    
    Put this up in the same way as iopoll is done for regular read/write IO.
    Make place for storing a cookie into struct io_uring_cmd on submission.
    Perform the completion using the ->uring_cmd_iopoll handler.
    
    Signed-off-by: Kanchan Joshi <joshi.k@samsung.com>
    Signed-off-by: Pankaj Raghav <p.raghav@samsung.com>
    Link: https://lore.kernel.org/r/20220823161443.49436-3-joshi.k@samsung.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:47 -04:00
Jeff Moyer bd9917d111 io_uring/opdef: rename SENDZC_NOTIF to SEND_ZC
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 9bd3f728223ebcfef8e9d087bdd142f0e388215d
Author: Stefan Metzmacher <metze@samba.org>
Date:   Fri Sep 16 23:36:25 2022 +0200

    io_uring/opdef: rename SENDZC_NOTIF to SEND_ZC
    
    It's confusing to see the string SENDZC_NOTIF in ftrace output
    when using IORING_OP_SEND_ZC.
    
    Fixes: b48c312be05e8 ("io_uring/net: simplify zerocopy send user API")
    Signed-off-by: Stefan Metzmacher <metze@samba.org>
    Cc: Pavel Begunkov <asml.silence@gmail.com>
    Cc: Jens Axboe <axboe@kernel.dk>
    Cc: io-uring@vger.kernel.org
    Reviewed-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/8e5cd8616919c92b6c3c7b6ea419fdffd5b97f3c.1663363798.git.metze@samba.org
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:23:01 -04:00
Jeff Moyer b2fa46daef io_uring/net: simplify zerocopy send user API
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit b48c312be05e83b55a4d58bf61f80b4a3288fb7e
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Thu Sep 1 11:54:04 2022 +0100

    io_uring/net: simplify zerocopy send user API
    
    Following user feedback, this patch simplifies zerocopy send API. One of
    the main complaints is that the current API is difficult with the
    userspace managing notification slots, and then send retries with error
    handling make it even worse.
    
    Instead of keeping notification slots change it to the per-request
    notifications model, which posts both completion and notification CQEs
    for each request when any data has been sent, and only one CQE if it
    fails. All notification CQEs will have IORING_CQE_F_NOTIF set and
    IORING_CQE_F_MORE in completion CQEs indicates whether to wait a
    notification or not.
    
    IOSQE_CQE_SKIP_SUCCESS is disallowed with zerocopy sends for now.
    
    This is less flexible, but greatly simplifies the user API and also the
    kernel implementation. We reuse notif helpers in this patch, but in the
    future there won't be need for keeping two requests.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/95287640ab98fc9417370afb16e310677c63e6ce.1662027856.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:22:58 -04:00
Jeff Moyer 37fde95f60 Revert "io_uring: rename IORING_OP_FILES_UPDATE"
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit d9808ceb3129b811becebdee3ec96d189c83e56c
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Thu Sep 1 11:54:02 2022 +0100

    Revert "io_uring: rename IORING_OP_FILES_UPDATE"
    
    This reverts commit 4379d5f15b3fd4224c37841029178aa8082a242e.
    
    We removed notification flushing, also cleanup uapi preparation changes
    to not pollute it.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/89edc3905350f91e1b6e26d9dbf42ee44fd451a2.1662027856.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:22:57 -04:00
Jeff Moyer adda922e38 io_uring/net: save address for sendzc async execution
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 581711c46612c1fd7f98960f9ad53f04fdb89853
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Wed Aug 24 13:07:43 2022 +0100

    io_uring/net: save address for sendzc async execution
    
    We usually copy all bits that a request needs from the userspace for
    async execution, so the userspace can keep them on the stack. However,
    send zerocopy violates this pattern for addresses and may reloads it
    e.g. from io-wq. Save the address if any in ->async_data as usual.
    
    Reported-by: Stefan Metzmacher <metze@samba.org>
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/d7512d7aa9abcd36e9afe1a4d292a24cb2d157e5.1661342812.git.asml.silence@gmail.com
    [axboe: fold in incremental fix]
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-05-05 15:22:55 -04:00
Jeff Moyer c447d13dbe io_uring: rename IORING_OP_FILES_UPDATE
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 4379d5f15b3fd4224c37841029178aa8082a242e
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Tue Jul 12 21:52:48 2022 +0100

    io_uring: rename IORING_OP_FILES_UPDATE
    
    IORING_OP_FILES_UPDATE will be a more generic opcode serving different
    resource types, rename it into IORING_OP_RSRC_UPDATE and add subtype
    handling.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/0a907133907d9af3415a8a7aa1802c6aa97c03c6.1657643355.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-04-29 08:26:02 -04:00
Jeff Moyer 7d666a408d io_uring: wire send zc request type
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 06a5464be84e4ae48394d34441baf34bf9706827
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Tue Jul 12 21:52:43 2022 +0100

    io_uring: wire send zc request type
    
    Add a new io_uring opcode IORING_OP_SENDZC. The main distinction from
    IORING_OP_SEND is that the user should specify a notification slot
    index in sqe::notification_idx and the buffers are safe to reuse only
    when the used notification is flushed and completes.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/a80387c6a68ce9cf99b3b6ef6f71068468761fb7.1657643355.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-04-29 08:21:02 -04:00
Jeff Moyer 8034d812f4 io_uring: kill extra io_uring_types.h includes
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit 27a9d66fec77cff0e32d2ecd5d0ac7ef878a7bb0
Author: Pavel Begunkov <asml.silence@gmail.com>
Date:   Thu Jun 16 13:57:18 2022 +0100

    io_uring: kill extra io_uring_types.h includes
    
    io_uring/io_uring.h already includes io_uring_types.h, no need to
    include it every time. Kill it in a bunch of places, it prepares us for
    following patches.
    
    Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
    Link: https://lore.kernel.org/r/94d8c943fbe0ef949981c508ddcee7fc1c18850f.1655384063.git.asml.silence@gmail.com
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-04-29 06:19:02 -04:00
Jeff Moyer 8b23020e0d io_uring: move opcode table to opdef.c
Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=2068237

commit d9b57aa3cfc792ccac6858376c017dbea6cb2872
Author: Jens Axboe <axboe@kernel.dk>
Date:   Wed Jun 15 16:27:42 2022 -0600

    io_uring: move opcode table to opdef.c
    
    We already have the declarations in opdef.h, move the rest into its own
    file rather than in the main io_uring.c file.
    
    Signed-off-by: Jens Axboe <axboe@kernel.dk>

Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
2023-04-29 05:42:02 -04:00