| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * FUSE inode io modes.
- *
- * Copyright (c) 2024 CTERA Networks.
- */
- #include "fuse_i.h"
- #include <linux/kernel.h>
- #include <linux/sched.h>
- #include <linux/file.h>
- #include <linux/fs.h>
- /*
- * Return true if need to wait for new opens in caching mode.
- */
- static inline bool fuse_is_io_cache_wait(struct fuse_inode *fi)
- {
- return READ_ONCE(fi->iocachectr) < 0 && !fuse_inode_backing(fi);
- }
- /*
- * Called on cached file open() and on first mmap() of direct_io file.
- * Takes cached_io inode mode reference to be dropped on file release.
- *
- * Blocks new parallel dio writes and waits for the in-progress parallel dio
- * writes to complete.
- */
- int fuse_file_cached_io_open(struct inode *inode, struct fuse_file *ff)
- {
- struct fuse_inode *fi = get_fuse_inode(inode);
- /* There are no io modes if server does not implement open */
- if (!ff->args)
- return 0;
- spin_lock(&fi->lock);
- /*
- * Setting the bit advises new direct-io writes to use an exclusive
- * lock - without it the wait below might be forever.
- */
- while (fuse_is_io_cache_wait(fi)) {
- set_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
- spin_unlock(&fi->lock);
- wait_event(fi->direct_io_waitq, !fuse_is_io_cache_wait(fi));
- spin_lock(&fi->lock);
- }
- /*
- * Check if inode entered passthrough io mode while waiting for parallel
- * dio write completion.
- */
- if (fuse_inode_backing(fi)) {
- clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
- spin_unlock(&fi->lock);
- return -ETXTBSY;
- }
- WARN_ON(ff->iomode == IOM_UNCACHED);
- if (ff->iomode == IOM_NONE) {
- ff->iomode = IOM_CACHED;
- if (fi->iocachectr == 0)
- set_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
- fi->iocachectr++;
- }
- spin_unlock(&fi->lock);
- return 0;
- }
- static void fuse_file_cached_io_release(struct fuse_file *ff,
- struct fuse_inode *fi)
- {
- spin_lock(&fi->lock);
- WARN_ON(fi->iocachectr <= 0);
- WARN_ON(ff->iomode != IOM_CACHED);
- ff->iomode = IOM_NONE;
- fi->iocachectr--;
- if (fi->iocachectr == 0)
- clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state);
- spin_unlock(&fi->lock);
- }
- /* Start strictly uncached io mode where cache access is not allowed */
- int fuse_inode_uncached_io_start(struct fuse_inode *fi, struct fuse_backing *fb)
- {
- struct fuse_backing *oldfb;
- int err = 0;
- spin_lock(&fi->lock);
- /* deny conflicting backing files on same fuse inode */
- oldfb = fuse_inode_backing(fi);
- if (fb && oldfb && oldfb != fb) {
- err = -EBUSY;
- goto unlock;
- }
- if (fi->iocachectr > 0) {
- err = -ETXTBSY;
- goto unlock;
- }
- fi->iocachectr--;
- /* fuse inode holds a single refcount of backing file */
- if (fb && !oldfb) {
- oldfb = fuse_inode_backing_set(fi, fb);
- WARN_ON_ONCE(oldfb != NULL);
- } else {
- fuse_backing_put(fb);
- }
- unlock:
- spin_unlock(&fi->lock);
- return err;
- }
- /* Takes uncached_io inode mode reference to be dropped on file release */
- static int fuse_file_uncached_io_open(struct inode *inode,
- struct fuse_file *ff,
- struct fuse_backing *fb)
- {
- struct fuse_inode *fi = get_fuse_inode(inode);
- int err;
- err = fuse_inode_uncached_io_start(fi, fb);
- if (err)
- return err;
- WARN_ON(ff->iomode != IOM_NONE);
- ff->iomode = IOM_UNCACHED;
- return 0;
- }
- void fuse_inode_uncached_io_end(struct fuse_inode *fi)
- {
- struct fuse_backing *oldfb = NULL;
- spin_lock(&fi->lock);
- WARN_ON(fi->iocachectr >= 0);
- fi->iocachectr++;
- if (!fi->iocachectr) {
- wake_up(&fi->direct_io_waitq);
- oldfb = fuse_inode_backing_set(fi, NULL);
- }
- spin_unlock(&fi->lock);
- if (oldfb)
- fuse_backing_put(oldfb);
- }
- /* Drop uncached_io reference from passthrough open */
- static void fuse_file_uncached_io_release(struct fuse_file *ff,
- struct fuse_inode *fi)
- {
- WARN_ON(ff->iomode != IOM_UNCACHED);
- ff->iomode = IOM_NONE;
- fuse_inode_uncached_io_end(fi);
- }
- /*
- * Open flags that are allowed in combination with FOPEN_PASSTHROUGH.
- * A combination of FOPEN_PASSTHROUGH and FOPEN_DIRECT_IO means that read/write
- * operations go directly to the server, but mmap is done on the backing file.
- * FOPEN_PASSTHROUGH mode should not co-exist with any users of the fuse inode
- * page cache, so FOPEN_KEEP_CACHE is a strange and undesired combination.
- */
- #define FOPEN_PASSTHROUGH_MASK \
- (FOPEN_PASSTHROUGH | FOPEN_DIRECT_IO | FOPEN_PARALLEL_DIRECT_WRITES | \
- FOPEN_NOFLUSH)
- static int fuse_file_passthrough_open(struct inode *inode, struct file *file)
- {
- struct fuse_file *ff = file->private_data;
- struct fuse_conn *fc = get_fuse_conn(inode);
- struct fuse_backing *fb;
- int err;
- /* Check allowed conditions for file open in passthrough mode */
- if (!IS_ENABLED(CONFIG_FUSE_PASSTHROUGH) || !fc->passthrough ||
- (ff->open_flags & ~FOPEN_PASSTHROUGH_MASK))
- return -EINVAL;
- fb = fuse_passthrough_open(file, inode,
- ff->args->open_outarg.backing_id);
- if (IS_ERR(fb))
- return PTR_ERR(fb);
- /* First passthrough file open denies caching inode io mode */
- err = fuse_file_uncached_io_open(inode, ff, fb);
- if (!err)
- return 0;
- fuse_passthrough_release(ff, fb);
- fuse_backing_put(fb);
- return err;
- }
- /* Request access to submit new io to inode via open file */
- int fuse_file_io_open(struct file *file, struct inode *inode)
- {
- struct fuse_file *ff = file->private_data;
- struct fuse_inode *fi = get_fuse_inode(inode);
- int err;
- /*
- * io modes are not relevant with DAX and with server that does not
- * implement open.
- */
- if (FUSE_IS_DAX(inode) || !ff->args)
- return 0;
- /*
- * Server is expected to use FOPEN_PASSTHROUGH for all opens of an inode
- * which is already open for passthrough.
- */
- err = -EINVAL;
- if (fuse_inode_backing(fi) && !(ff->open_flags & FOPEN_PASSTHROUGH))
- goto fail;
- /*
- * FOPEN_PARALLEL_DIRECT_WRITES requires FOPEN_DIRECT_IO.
- */
- if (!(ff->open_flags & FOPEN_DIRECT_IO))
- ff->open_flags &= ~FOPEN_PARALLEL_DIRECT_WRITES;
- /*
- * First passthrough file open denies caching inode io mode.
- * First caching file open enters caching inode io mode.
- *
- * Note that if user opens a file open with O_DIRECT, but server did
- * not specify FOPEN_DIRECT_IO, a later fcntl() could remove O_DIRECT,
- * so we put the inode in caching mode to prevent parallel dio.
- */
- if ((ff->open_flags & FOPEN_DIRECT_IO) &&
- !(ff->open_flags & FOPEN_PASSTHROUGH))
- return 0;
- if (ff->open_flags & FOPEN_PASSTHROUGH)
- err = fuse_file_passthrough_open(inode, file);
- else
- err = fuse_file_cached_io_open(inode, ff);
- if (err)
- goto fail;
- return 0;
- fail:
- pr_debug("failed to open file in requested io mode (open_flags=0x%x, err=%i).\n",
- ff->open_flags, err);
- /*
- * The file open mode determines the inode io mode.
- * Using incorrect open mode is a server mistake, which results in
- * user visible failure of open() with EIO error.
- */
- return -EIO;
- }
- /* No more pending io and no new io possible to inode via open/mmapped file */
- void fuse_file_io_release(struct fuse_file *ff, struct inode *inode)
- {
- struct fuse_inode *fi = get_fuse_inode(inode);
- /*
- * Last passthrough file close allows caching inode io mode.
- * Last caching file close exits caching inode io mode.
- */
- switch (ff->iomode) {
- case IOM_NONE:
- /* Nothing to do */
- break;
- case IOM_UNCACHED:
- fuse_file_uncached_io_release(ff, fi);
- break;
- case IOM_CACHED:
- fuse_file_cached_io_release(ff, fi);
- break;
- }
- }
|