|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * FUSE inode io modes. | 
|  | * | 
|  | * Copyright (c) 2024 CTERA Networks. | 
|  | */ | 
|  |  | 
|  | #include "fuse_i.h" | 
|  |  | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/sched.h> | 
|  | #include <linux/file.h> | 
|  | #include <linux/fs.h> | 
|  |  | 
|  | /* | 
|  | * Return true if need to wait for new opens in caching mode. | 
|  | */ | 
|  | static inline bool fuse_is_io_cache_wait(struct fuse_inode *fi) | 
|  | { | 
|  | return READ_ONCE(fi->iocachectr) < 0 && !fuse_inode_backing(fi); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Called on cached file open() and on first mmap() of direct_io file. | 
|  | * Takes cached_io inode mode reference to be dropped on file release. | 
|  | * | 
|  | * Blocks new parallel dio writes and waits for the in-progress parallel dio | 
|  | * writes to complete. | 
|  | */ | 
|  | int fuse_file_cached_io_open(struct inode *inode, struct fuse_file *ff) | 
|  | { | 
|  | struct fuse_inode *fi = get_fuse_inode(inode); | 
|  |  | 
|  | /* There are no io modes if server does not implement open */ | 
|  | if (!ff->args) | 
|  | return 0; | 
|  |  | 
|  | spin_lock(&fi->lock); | 
|  | /* | 
|  | * Setting the bit advises new direct-io writes to use an exclusive | 
|  | * lock - without it the wait below might be forever. | 
|  | */ | 
|  | while (fuse_is_io_cache_wait(fi)) { | 
|  | set_bit(FUSE_I_CACHE_IO_MODE, &fi->state); | 
|  | spin_unlock(&fi->lock); | 
|  | wait_event(fi->direct_io_waitq, !fuse_is_io_cache_wait(fi)); | 
|  | spin_lock(&fi->lock); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Check if inode entered passthrough io mode while waiting for parallel | 
|  | * dio write completion. | 
|  | */ | 
|  | if (fuse_inode_backing(fi)) { | 
|  | clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state); | 
|  | spin_unlock(&fi->lock); | 
|  | return -ETXTBSY; | 
|  | } | 
|  |  | 
|  | WARN_ON(ff->iomode == IOM_UNCACHED); | 
|  | if (ff->iomode == IOM_NONE) { | 
|  | ff->iomode = IOM_CACHED; | 
|  | if (fi->iocachectr == 0) | 
|  | set_bit(FUSE_I_CACHE_IO_MODE, &fi->state); | 
|  | fi->iocachectr++; | 
|  | } | 
|  | spin_unlock(&fi->lock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void fuse_file_cached_io_release(struct fuse_file *ff, | 
|  | struct fuse_inode *fi) | 
|  | { | 
|  | spin_lock(&fi->lock); | 
|  | WARN_ON(fi->iocachectr <= 0); | 
|  | WARN_ON(ff->iomode != IOM_CACHED); | 
|  | ff->iomode = IOM_NONE; | 
|  | fi->iocachectr--; | 
|  | if (fi->iocachectr == 0) | 
|  | clear_bit(FUSE_I_CACHE_IO_MODE, &fi->state); | 
|  | spin_unlock(&fi->lock); | 
|  | } | 
|  |  | 
|  | /* Start strictly uncached io mode where cache access is not allowed */ | 
|  | int fuse_inode_uncached_io_start(struct fuse_inode *fi, struct fuse_backing *fb) | 
|  | { | 
|  | struct fuse_backing *oldfb; | 
|  | int err = 0; | 
|  |  | 
|  | spin_lock(&fi->lock); | 
|  | /* deny conflicting backing files on same fuse inode */ | 
|  | oldfb = fuse_inode_backing(fi); | 
|  | if (fb && oldfb && oldfb != fb) { | 
|  | err = -EBUSY; | 
|  | goto unlock; | 
|  | } | 
|  | if (fi->iocachectr > 0) { | 
|  | err = -ETXTBSY; | 
|  | goto unlock; | 
|  | } | 
|  | fi->iocachectr--; | 
|  |  | 
|  | /* fuse inode holds a single refcount of backing file */ | 
|  | if (fb && !oldfb) { | 
|  | oldfb = fuse_inode_backing_set(fi, fb); | 
|  | WARN_ON_ONCE(oldfb != NULL); | 
|  | } else { | 
|  | fuse_backing_put(fb); | 
|  | } | 
|  | unlock: | 
|  | spin_unlock(&fi->lock); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | /* Takes uncached_io inode mode reference to be dropped on file release */ | 
|  | static int fuse_file_uncached_io_open(struct inode *inode, | 
|  | struct fuse_file *ff, | 
|  | struct fuse_backing *fb) | 
|  | { | 
|  | struct fuse_inode *fi = get_fuse_inode(inode); | 
|  | int err; | 
|  |  | 
|  | err = fuse_inode_uncached_io_start(fi, fb); | 
|  | if (err) | 
|  | return err; | 
|  |  | 
|  | WARN_ON(ff->iomode != IOM_NONE); | 
|  | ff->iomode = IOM_UNCACHED; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | void fuse_inode_uncached_io_end(struct fuse_inode *fi) | 
|  | { | 
|  | struct fuse_backing *oldfb = NULL; | 
|  |  | 
|  | spin_lock(&fi->lock); | 
|  | WARN_ON(fi->iocachectr >= 0); | 
|  | fi->iocachectr++; | 
|  | if (!fi->iocachectr) { | 
|  | wake_up(&fi->direct_io_waitq); | 
|  | oldfb = fuse_inode_backing_set(fi, NULL); | 
|  | } | 
|  | spin_unlock(&fi->lock); | 
|  | if (oldfb) | 
|  | fuse_backing_put(oldfb); | 
|  | } | 
|  |  | 
|  | /* Drop uncached_io reference from passthrough open */ | 
|  | static void fuse_file_uncached_io_release(struct fuse_file *ff, | 
|  | struct fuse_inode *fi) | 
|  | { | 
|  | WARN_ON(ff->iomode != IOM_UNCACHED); | 
|  | ff->iomode = IOM_NONE; | 
|  | fuse_inode_uncached_io_end(fi); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Open flags that are allowed in combination with FOPEN_PASSTHROUGH. | 
|  | * A combination of FOPEN_PASSTHROUGH and FOPEN_DIRECT_IO means that read/write | 
|  | * operations go directly to the server, but mmap is done on the backing file. | 
|  | * FOPEN_PASSTHROUGH mode should not co-exist with any users of the fuse inode | 
|  | * page cache, so FOPEN_KEEP_CACHE is a strange and undesired combination. | 
|  | */ | 
|  | #define FOPEN_PASSTHROUGH_MASK \ | 
|  | (FOPEN_PASSTHROUGH | FOPEN_DIRECT_IO | FOPEN_PARALLEL_DIRECT_WRITES | \ | 
|  | FOPEN_NOFLUSH) | 
|  |  | 
|  | static int fuse_file_passthrough_open(struct inode *inode, struct file *file) | 
|  | { | 
|  | struct fuse_file *ff = file->private_data; | 
|  | struct fuse_conn *fc = get_fuse_conn(inode); | 
|  | struct fuse_backing *fb; | 
|  | int err; | 
|  |  | 
|  | /* Check allowed conditions for file open in passthrough mode */ | 
|  | if (!IS_ENABLED(CONFIG_FUSE_PASSTHROUGH) || !fc->passthrough || | 
|  | (ff->open_flags & ~FOPEN_PASSTHROUGH_MASK)) | 
|  | return -EINVAL; | 
|  |  | 
|  | fb = fuse_passthrough_open(file, ff->args->open_outarg.backing_id); | 
|  | if (IS_ERR(fb)) | 
|  | return PTR_ERR(fb); | 
|  |  | 
|  | /* First passthrough file open denies caching inode io mode */ | 
|  | err = fuse_file_uncached_io_open(inode, ff, fb); | 
|  | if (!err) | 
|  | return 0; | 
|  |  | 
|  | fuse_passthrough_release(ff, fb); | 
|  | fuse_backing_put(fb); | 
|  |  | 
|  | return err; | 
|  | } | 
|  |  | 
|  | /* Request access to submit new io to inode via open file */ | 
|  | int fuse_file_io_open(struct file *file, struct inode *inode) | 
|  | { | 
|  | struct fuse_file *ff = file->private_data; | 
|  | struct fuse_inode *fi = get_fuse_inode(inode); | 
|  | int err; | 
|  |  | 
|  | /* | 
|  | * io modes are not relevant with DAX and with server that does not | 
|  | * implement open. | 
|  | */ | 
|  | if (FUSE_IS_DAX(inode) || !ff->args) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * Server is expected to use FOPEN_PASSTHROUGH for all opens of an inode | 
|  | * which is already open for passthrough. | 
|  | */ | 
|  | err = -EINVAL; | 
|  | if (fuse_inode_backing(fi) && !(ff->open_flags & FOPEN_PASSTHROUGH)) | 
|  | goto fail; | 
|  |  | 
|  | /* | 
|  | * FOPEN_PARALLEL_DIRECT_WRITES requires FOPEN_DIRECT_IO. | 
|  | */ | 
|  | if (!(ff->open_flags & FOPEN_DIRECT_IO)) | 
|  | ff->open_flags &= ~FOPEN_PARALLEL_DIRECT_WRITES; | 
|  |  | 
|  | /* | 
|  | * First passthrough file open denies caching inode io mode. | 
|  | * First caching file open enters caching inode io mode. | 
|  | * | 
|  | * Note that if user opens a file open with O_DIRECT, but server did | 
|  | * not specify FOPEN_DIRECT_IO, a later fcntl() could remove O_DIRECT, | 
|  | * so we put the inode in caching mode to prevent parallel dio. | 
|  | */ | 
|  | if ((ff->open_flags & FOPEN_DIRECT_IO) && | 
|  | !(ff->open_flags & FOPEN_PASSTHROUGH)) | 
|  | return 0; | 
|  |  | 
|  | if (ff->open_flags & FOPEN_PASSTHROUGH) | 
|  | err = fuse_file_passthrough_open(inode, file); | 
|  | else | 
|  | err = fuse_file_cached_io_open(inode, ff); | 
|  | if (err) | 
|  | goto fail; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | fail: | 
|  | pr_debug("failed to open file in requested io mode (open_flags=0x%x, err=%i).\n", | 
|  | ff->open_flags, err); | 
|  | /* | 
|  | * The file open mode determines the inode io mode. | 
|  | * Using incorrect open mode is a server mistake, which results in | 
|  | * user visible failure of open() with EIO error. | 
|  | */ | 
|  | return -EIO; | 
|  | } | 
|  |  | 
|  | /* No more pending io and no new io possible to inode via open/mmapped file */ | 
|  | void fuse_file_io_release(struct fuse_file *ff, struct inode *inode) | 
|  | { | 
|  | struct fuse_inode *fi = get_fuse_inode(inode); | 
|  |  | 
|  | /* | 
|  | * Last passthrough file close allows caching inode io mode. | 
|  | * Last caching file close exits caching inode io mode. | 
|  | */ | 
|  | switch (ff->iomode) { | 
|  | case IOM_NONE: | 
|  | /* Nothing to do */ | 
|  | break; | 
|  | case IOM_UNCACHED: | 
|  | fuse_file_uncached_io_release(ff, fi); | 
|  | break; | 
|  | case IOM_CACHED: | 
|  | fuse_file_cached_io_release(ff, fi); | 
|  | break; | 
|  | } | 
|  | } |