On Mon, Oct 21, 2019 at 11:58:06AM +0100, Dr. David Alan Gilbert (git) wrote: > From: "Dr. David Alan Gilbert" <dgilb...@redhat.com> > > fuse_lowlevel is one of the largest files from the library > and does most of the work. Add it separately to keep the diff > sizes small. > Again this is from upstream a0d713619d4f21e8240ede38eefe6101ea580724 > > Signed-off-by: Dr. David Alan Gilbert <dgilb...@redhat.com> > --- > contrib/virtiofsd/fuse_lowlevel.c | 3104 +++++++++++++++++++++++++++++ > 1 file changed, 3104 insertions(+) > create mode 100644 contrib/virtiofsd/fuse_lowlevel.c > > diff --git a/contrib/virtiofsd/fuse_lowlevel.c > b/contrib/virtiofsd/fuse_lowlevel.c > new file mode 100644 > index 0000000000..f7fbc8f5db > --- /dev/null > +++ b/contrib/virtiofsd/fuse_lowlevel.c > @@ -0,0 +1,3104 @@ > +/* > + FUSE: Filesystem in Userspace > + Copyright (C) 2001-2007 Miklos Szeredi <mik...@szeredi.hu> > + > + Implementation of (most of) the low-level FUSE API. The session loop > + functions are implemented in separate files. > + > + This program can be distributed under the terms of the GNU LGPLv2. > + See the file COPYING.LIB
I guess we need COPYING.LIB then? > +*/ > + > +#define _GNU_SOURCE > + > +#include "config.h" > +#include "fuse_i.h" > +#include "fuse_kernel.h" > +#include "fuse_opt.h" > +#include "fuse_misc.h" > +#include "mount_util.h" > + > +#include <stdio.h> > +#include <stdlib.h> > +#include <stddef.h> > +#include <string.h> > +#include <unistd.h> > +#include <limits.h> > +#include <errno.h> > +#include <assert.h> > +#include <sys/file.h> > + > +#ifndef F_LINUX_SPECIFIC_BASE > +#define F_LINUX_SPECIFIC_BASE 1024 > +#endif > +#ifndef F_SETPIPE_SZ > +#define F_SETPIPE_SZ (F_LINUX_SPECIFIC_BASE + 7) > +#endif > + > + > +#define PARAM(inarg) (((char *)(inarg)) + sizeof(*(inarg))) > +#define OFFSET_MAX 0x7fffffffffffffffLL > + > +#define container_of(ptr, type, member) ({ \ > + const typeof( ((type *)0)->member ) *__mptr = (ptr); \ > + (type *)( (char *)__mptr - offsetof(type,member) );}) > + > +struct fuse_pollhandle { > + uint64_t kh; > + struct fuse_session *se; > +}; > + > +static size_t pagesize; > + > +static __attribute__((constructor)) void fuse_ll_init_pagesize(void) > +{ > + pagesize = getpagesize(); > +} > + > +static void convert_stat(const struct stat *stbuf, struct fuse_attr *attr) > +{ > + attr->ino = stbuf->st_ino; > + attr->mode = stbuf->st_mode; > + attr->nlink = stbuf->st_nlink; > + attr->uid = stbuf->st_uid; > + attr->gid = stbuf->st_gid; > + attr->rdev = stbuf->st_rdev; > + attr->size = stbuf->st_size; > + attr->blksize = stbuf->st_blksize; > + attr->blocks = stbuf->st_blocks; > + attr->atime = stbuf->st_atime; > + attr->mtime = stbuf->st_mtime; > + attr->ctime = stbuf->st_ctime; > + attr->atimensec = ST_ATIM_NSEC(stbuf); > + attr->mtimensec = ST_MTIM_NSEC(stbuf); > + attr->ctimensec = ST_CTIM_NSEC(stbuf); > +} > + > +static void convert_attr(const struct fuse_setattr_in *attr, struct stat > *stbuf) > +{ > + stbuf->st_mode = attr->mode; > + stbuf->st_uid = attr->uid; > + stbuf->st_gid = attr->gid; > + stbuf->st_size = attr->size; > + stbuf->st_atime = attr->atime; > + stbuf->st_mtime = attr->mtime; > + stbuf->st_ctime = attr->ctime; > + ST_ATIM_NSEC_SET(stbuf, attr->atimensec); > + ST_MTIM_NSEC_SET(stbuf, attr->mtimensec); > + ST_CTIM_NSEC_SET(stbuf, attr->ctimensec); > +} > + > +static size_t iov_length(const struct iovec *iov, size_t count) > +{ > + size_t seg; > + size_t ret = 0; > + > + for (seg = 0; seg < count; seg++) > + ret += iov[seg].iov_len; > + return ret; > +} > + > +static void list_init_req(struct fuse_req *req) > +{ > + req->next = req; > + req->prev = req; > +} > + > +static void list_del_req(struct fuse_req *req) > +{ > + struct fuse_req *prev = req->prev; > + struct fuse_req *next = req->next; > + prev->next = next; > + next->prev = prev; > +} > + > +static void list_add_req(struct fuse_req *req, struct fuse_req *next) > +{ > + struct fuse_req *prev = next->prev; > + req->next = next; > + req->prev = prev; > + prev->next = req; > + next->prev = req; > +} > + > +static void destroy_req(fuse_req_t req) > +{ > + pthread_mutex_destroy(&req->lock); > + free(req); > +} > + > +void fuse_free_req(fuse_req_t req) > +{ > + int ctr; > + struct fuse_session *se = req->se; > + > + pthread_mutex_lock(&se->lock); > + req->u.ni.func = NULL; > + req->u.ni.data = NULL; > + list_del_req(req); > + ctr = --req->ctr; > + fuse_chan_put(req->ch); > + req->ch = NULL; > + pthread_mutex_unlock(&se->lock); > + if (!ctr) > + destroy_req(req); > +} > + > +static struct fuse_req *fuse_ll_alloc_req(struct fuse_session *se) > +{ > + struct fuse_req *req; > + > + req = (struct fuse_req *) calloc(1, sizeof(struct fuse_req)); > + if (req == NULL) { > + fuse_log(FUSE_LOG_ERR, "fuse: failed to allocate request\n"); > + } else { > + req->se = se; > + req->ctr = 1; > + list_init_req(req); > + fuse_mutex_init(&req->lock); > + } > + > + return req; > +} > + > +/* Send data. If *ch* is NULL, send via session master fd */ > +static int fuse_send_msg(struct fuse_session *se, struct fuse_chan *ch, > + struct iovec *iov, int count) > +{ > + struct fuse_out_header *out = iov[0].iov_base; > + > + out->len = iov_length(iov, count); > + if (se->debug) { > + if (out->unique == 0) { > + fuse_log(FUSE_LOG_DEBUG, "NOTIFY: code=%d length=%u\n", > + out->error, out->len); > + } else if (out->error) { > + fuse_log(FUSE_LOG_DEBUG, > + " unique: %llu, error: %i (%s), outsize: > %i\n", > + (unsigned long long) out->unique, out->error, > + strerror(-out->error), out->len); > + } else { > + fuse_log(FUSE_LOG_DEBUG, > + " unique: %llu, success, outsize: %i\n", > + (unsigned long long) out->unique, out->len); > + } > + } > + > + ssize_t res = writev(ch ? ch->fd : se->fd, > + iov, count); > + int err = errno; > + > + if (res == -1) { > + assert(se != NULL); > + > + /* ENOENT means the operation was interrupted */ > + if (!fuse_session_exited(se) && err != ENOENT) > + perror("fuse: writing device"); > + return -err; > + } > + > + return 0; > +} > + > + > +int fuse_send_reply_iov_nofree(fuse_req_t req, int error, struct iovec *iov, > + int count) > +{ > + struct fuse_out_header out; > + > + if (error <= -1000 || error > 0) { > + fuse_log(FUSE_LOG_ERR, "fuse: bad error value: %i\n", error); > + error = -ERANGE; > + } > + > + out.unique = req->unique; > + out.error = error; > + > + iov[0].iov_base = &out; > + iov[0].iov_len = sizeof(struct fuse_out_header); > + > + return fuse_send_msg(req->se, req->ch, iov, count); > +} > + > +static int send_reply_iov(fuse_req_t req, int error, struct iovec *iov, > + int count) > +{ > + int res; > + > + res = fuse_send_reply_iov_nofree(req, error, iov, count); > + fuse_free_req(req); > + return res; > +} > + > +static int send_reply(fuse_req_t req, int error, const void *arg, > + size_t argsize) > +{ > + struct iovec iov[2]; > + int count = 1; > + if (argsize) { > + iov[1].iov_base = (void *) arg; > + iov[1].iov_len = argsize; > + count++; > + } > + return send_reply_iov(req, error, iov, count); > +} > + > +int fuse_reply_iov(fuse_req_t req, const struct iovec *iov, int count) > +{ > + int res; > + struct iovec *padded_iov; > + > + padded_iov = malloc((count + 1) * sizeof(struct iovec)); > + if (padded_iov == NULL) > + return fuse_reply_err(req, ENOMEM); > + > + memcpy(padded_iov + 1, iov, count * sizeof(struct iovec)); > + count++; > + > + res = send_reply_iov(req, 0, padded_iov, count); > + free(padded_iov); > + > + return res; > +} > + > + > +/* `buf` is allowed to be empty so that the proper size may be > + allocated by the caller */ > +size_t fuse_add_direntry(fuse_req_t req, char *buf, size_t bufsize, > + const char *name, const struct stat *stbuf, off_t off) > +{ > + (void)req; > + size_t namelen; > + size_t entlen; > + size_t entlen_padded; > + struct fuse_dirent *dirent; > + > + namelen = strlen(name); > + entlen = FUSE_NAME_OFFSET + namelen; > + entlen_padded = FUSE_DIRENT_ALIGN(entlen); > + > + if ((buf == NULL) || (entlen_padded > bufsize)) > + return entlen_padded; > + > + dirent = (struct fuse_dirent*) buf; > + dirent->ino = stbuf->st_ino; > + dirent->off = off; > + dirent->namelen = namelen; > + dirent->type = (stbuf->st_mode & S_IFMT) >> 12; > + memcpy(dirent->name, name, namelen); > + memset(dirent->name + namelen, 0, entlen_padded - entlen); > + > + return entlen_padded; > +} > + > +static void convert_statfs(const struct statvfs *stbuf, > + struct fuse_kstatfs *kstatfs) > +{ > + kstatfs->bsize = stbuf->f_bsize; > + kstatfs->frsize = stbuf->f_frsize; > + kstatfs->blocks = stbuf->f_blocks; > + kstatfs->bfree = stbuf->f_bfree; > + kstatfs->bavail = stbuf->f_bavail; > + kstatfs->files = stbuf->f_files; > + kstatfs->ffree = stbuf->f_ffree; > + kstatfs->namelen = stbuf->f_namemax; > +} > + > +static int send_reply_ok(fuse_req_t req, const void *arg, size_t argsize) > +{ > + return send_reply(req, 0, arg, argsize); > +} > + > +int fuse_reply_err(fuse_req_t req, int err) > +{ > + return send_reply(req, -err, NULL, 0); > +} > + > +void fuse_reply_none(fuse_req_t req) > +{ > + fuse_free_req(req); > +} > + > +static unsigned long calc_timeout_sec(double t) > +{ > + if (t > (double) ULONG_MAX) > + return ULONG_MAX; > + else if (t < 0.0) > + return 0; > + else > + return (unsigned long) t; > +} > + > +static unsigned int calc_timeout_nsec(double t) > +{ > + double f = t - (double) calc_timeout_sec(t); > + if (f < 0.0) > + return 0; > + else if (f >= 0.999999999) > + return 999999999; > + else > + return (unsigned int) (f * 1.0e9); > +} > + > +static void fill_entry(struct fuse_entry_out *arg, > + const struct fuse_entry_param *e) > +{ > + arg->nodeid = e->ino; > + arg->generation = e->generation; > + arg->entry_valid = calc_timeout_sec(e->entry_timeout); > + arg->entry_valid_nsec = calc_timeout_nsec(e->entry_timeout); > + arg->attr_valid = calc_timeout_sec(e->attr_timeout); > + arg->attr_valid_nsec = calc_timeout_nsec(e->attr_timeout); > + convert_stat(&e->attr, &arg->attr); > +} > + > +/* `buf` is allowed to be empty so that the proper size may be > + allocated by the caller */ > +size_t fuse_add_direntry_plus(fuse_req_t req, char *buf, size_t bufsize, > + const char *name, > + const struct fuse_entry_param *e, off_t off) > +{ > + (void)req; > + size_t namelen; > + size_t entlen; > + size_t entlen_padded; > + > + namelen = strlen(name); > + entlen = FUSE_NAME_OFFSET_DIRENTPLUS + namelen; > + entlen_padded = FUSE_DIRENT_ALIGN(entlen); > + if ((buf == NULL) || (entlen_padded > bufsize)) > + return entlen_padded; > + > + struct fuse_direntplus *dp = (struct fuse_direntplus *) buf; > + memset(&dp->entry_out, 0, sizeof(dp->entry_out)); > + fill_entry(&dp->entry_out, e); > + > + struct fuse_dirent *dirent = &dp->dirent; > + dirent->ino = e->attr.st_ino; > + dirent->off = off; > + dirent->namelen = namelen; > + dirent->type = (e->attr.st_mode & S_IFMT) >> 12; > + memcpy(dirent->name, name, namelen); > + memset(dirent->name + namelen, 0, entlen_padded - entlen); > + > + return entlen_padded; > +} > + > +static void fill_open(struct fuse_open_out *arg, > + const struct fuse_file_info *f) > +{ > + arg->fh = f->fh; > + if (f->direct_io) > + arg->open_flags |= FOPEN_DIRECT_IO; > + if (f->keep_cache) > + arg->open_flags |= FOPEN_KEEP_CACHE; > + if (f->cache_readdir) > + arg->open_flags |= FOPEN_CACHE_DIR; > + if (f->nonseekable) > + arg->open_flags |= FOPEN_NONSEEKABLE; > +} > + > +int fuse_reply_entry(fuse_req_t req, const struct fuse_entry_param *e) > +{ > + struct fuse_entry_out arg; > + size_t size = req->se->conn.proto_minor < 9 ? > + FUSE_COMPAT_ENTRY_OUT_SIZE : sizeof(arg); > + > + /* before ABI 7.4 e->ino == 0 was invalid, only ENOENT meant > + negative entry */ > + if (!e->ino && req->se->conn.proto_minor < 4) > + return fuse_reply_err(req, ENOENT); > + > + memset(&arg, 0, sizeof(arg)); > + fill_entry(&arg, e); > + return send_reply_ok(req, &arg, size); > +} > + > +int fuse_reply_create(fuse_req_t req, const struct fuse_entry_param *e, > + const struct fuse_file_info *f) > +{ > + char buf[sizeof(struct fuse_entry_out) + sizeof(struct fuse_open_out)]; > + size_t entrysize = req->se->conn.proto_minor < 9 ? > + FUSE_COMPAT_ENTRY_OUT_SIZE : sizeof(struct fuse_entry_out); > + struct fuse_entry_out *earg = (struct fuse_entry_out *) buf; > + struct fuse_open_out *oarg = (struct fuse_open_out *) (buf + entrysize); > + > + memset(buf, 0, sizeof(buf)); > + fill_entry(earg, e); > + fill_open(oarg, f); > + return send_reply_ok(req, buf, > + entrysize + sizeof(struct fuse_open_out)); > +} > + > +int fuse_reply_attr(fuse_req_t req, const struct stat *attr, > + double attr_timeout) > +{ > + struct fuse_attr_out arg; > + size_t size = req->se->conn.proto_minor < 9 ? > + FUSE_COMPAT_ATTR_OUT_SIZE : sizeof(arg); > + > + memset(&arg, 0, sizeof(arg)); > + arg.attr_valid = calc_timeout_sec(attr_timeout); > + arg.attr_valid_nsec = calc_timeout_nsec(attr_timeout); > + convert_stat(attr, &arg.attr); > + > + return send_reply_ok(req, &arg, size); > +} > + > +int fuse_reply_readlink(fuse_req_t req, const char *linkname) > +{ > + return send_reply_ok(req, linkname, strlen(linkname)); > +} > + > +int fuse_reply_open(fuse_req_t req, const struct fuse_file_info *f) > +{ > + struct fuse_open_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + fill_open(&arg, f); > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +int fuse_reply_write(fuse_req_t req, size_t count) > +{ > + struct fuse_write_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + arg.size = count; > + > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +int fuse_reply_buf(fuse_req_t req, const char *buf, size_t size) > +{ > + return send_reply_ok(req, buf, size); > +} > + > +static int fuse_send_data_iov_fallback(struct fuse_session *se, > + struct fuse_chan *ch, > + struct iovec *iov, int iov_count, > + struct fuse_bufvec *buf, > + size_t len) > +{ > + struct fuse_bufvec mem_buf = FUSE_BUFVEC_INIT(len); > + void *mbuf; > + int res; > + > + /* Optimize common case */ > + if (buf->count == 1 && buf->idx == 0 && buf->off == 0 && > + !(buf->buf[0].flags & FUSE_BUF_IS_FD)) { > + /* FIXME: also avoid memory copy if there are multiple buffers > + but none of them contain an fd */ > + > + iov[iov_count].iov_base = buf->buf[0].mem; > + iov[iov_count].iov_len = len; > + iov_count++; > + return fuse_send_msg(se, ch, iov, iov_count); > + } > + > + res = posix_memalign(&mbuf, pagesize, len); > + if (res != 0) > + return res; > + > + mem_buf.buf[0].mem = mbuf; > + res = fuse_buf_copy(&mem_buf, buf, 0); > + if (res < 0) { > + free(mbuf); > + return -res; > + } > + len = res; > + > + iov[iov_count].iov_base = mbuf; > + iov[iov_count].iov_len = len; > + iov_count++; > + res = fuse_send_msg(se, ch, iov, iov_count); > + free(mbuf); > + > + return res; > +} > + > +struct fuse_ll_pipe { > + size_t size; > + int can_grow; > + int pipe[2]; > +}; > + > +static void fuse_ll_pipe_free(struct fuse_ll_pipe *llp) > +{ > + close(llp->pipe[0]); > + close(llp->pipe[1]); > + free(llp); > +} > + > +#ifdef HAVE_SPLICE > +#if !defined(HAVE_PIPE2) || !defined(O_CLOEXEC) > +static int fuse_pipe(int fds[2]) > +{ > + int rv = pipe(fds); > + > + if (rv == -1) > + return rv; > + > + if (fcntl(fds[0], F_SETFL, O_NONBLOCK) == -1 || > + fcntl(fds[1], F_SETFL, O_NONBLOCK) == -1 || > + fcntl(fds[0], F_SETFD, FD_CLOEXEC) == -1 || > + fcntl(fds[1], F_SETFD, FD_CLOEXEC) == -1) { > + close(fds[0]); > + close(fds[1]); > + rv = -1; > + } > + return rv; > +} > +#else > +static int fuse_pipe(int fds[2]) > +{ > + return pipe2(fds, O_CLOEXEC | O_NONBLOCK); > +} > +#endif > + > +static struct fuse_ll_pipe *fuse_ll_get_pipe(struct fuse_session *se) > +{ > + struct fuse_ll_pipe *llp = pthread_getspecific(se->pipe_key); > + if (llp == NULL) { > + int res; > + > + llp = malloc(sizeof(struct fuse_ll_pipe)); > + if (llp == NULL) > + return NULL; > + > + res = fuse_pipe(llp->pipe); > + if (res == -1) { > + free(llp); > + return NULL; > + } > + > + /* > + *the default size is 16 pages on linux > + */ > + llp->size = pagesize * 16; > + llp->can_grow = 1; > + > + pthread_setspecific(se->pipe_key, llp); > + } > + > + return llp; > +} > +#endif > + > +static void fuse_ll_clear_pipe(struct fuse_session *se) > +{ > + struct fuse_ll_pipe *llp = pthread_getspecific(se->pipe_key); > + if (llp) { > + pthread_setspecific(se->pipe_key, NULL); > + fuse_ll_pipe_free(llp); > + } > +} > + > +#if defined(HAVE_SPLICE) && defined(HAVE_VMSPLICE) > +static int read_back(int fd, char *buf, size_t len) > +{ > + int res; > + > + res = read(fd, buf, len); > + if (res == -1) { > + fuse_log(FUSE_LOG_ERR, "fuse: internal error: failed to read > back from pipe: %s\n", strerror(errno)); > + return -EIO; > + } > + if (res != len) { > + fuse_log(FUSE_LOG_ERR, "fuse: internal error: short read back > from pipe: %i from %zi\n", res, len); > + return -EIO; > + } > + return 0; > +} > + > +static int grow_pipe_to_max(int pipefd) > +{ > + int max; > + int res; > + int maxfd; > + char buf[32]; > + > + maxfd = open("/proc/sys/fs/pipe-max-size", O_RDONLY); > + if (maxfd < 0) > + return -errno; > + > + res = read(maxfd, buf, sizeof(buf) - 1); > + if (res < 0) { > + int saved_errno; > + > + saved_errno = errno; > + close(maxfd); > + return -saved_errno; > + } > + close(maxfd); > + buf[res] = '\0'; > + > + max = atoi(buf); > + res = fcntl(pipefd, F_SETPIPE_SZ, max); > + if (res < 0) > + return -errno; > + return max; > +} > + > +static int fuse_send_data_iov(struct fuse_session *se, struct fuse_chan *ch, > + struct iovec *iov, int iov_count, > + struct fuse_bufvec *buf, unsigned int flags) > +{ > + int res; > + size_t len = fuse_buf_size(buf); > + struct fuse_out_header *out = iov[0].iov_base; > + struct fuse_ll_pipe *llp; > + int splice_flags; > + size_t pipesize; > + size_t total_fd_size; > + size_t idx; > + size_t headerlen; > + struct fuse_bufvec pipe_buf = FUSE_BUFVEC_INIT(len); > + > + if (se->broken_splice_nonblock) > + goto fallback; > + > + if (flags & FUSE_BUF_NO_SPLICE) > + goto fallback; > + > + total_fd_size = 0; > + for (idx = buf->idx; idx < buf->count; idx++) { > + if (buf->buf[idx].flags & FUSE_BUF_IS_FD) { > + total_fd_size = buf->buf[idx].size; > + if (idx == buf->idx) > + total_fd_size -= buf->off; > + } > + } > + if (total_fd_size < 2 * pagesize) > + goto fallback; > + > + if (se->conn.proto_minor < 14 || > + !(se->conn.want & FUSE_CAP_SPLICE_WRITE)) > + goto fallback; > + > + llp = fuse_ll_get_pipe(se); > + if (llp == NULL) > + goto fallback; > + > + > + headerlen = iov_length(iov, iov_count); > + > + out->len = headerlen + len; > + > + /* > + * Heuristic for the required pipe size, does not work if the > + * source contains less than page size fragments > + */ > + pipesize = pagesize * (iov_count + buf->count + 1) + out->len; > + > + if (llp->size < pipesize) { > + if (llp->can_grow) { > + res = fcntl(llp->pipe[0], F_SETPIPE_SZ, pipesize); > + if (res == -1) { > + res = grow_pipe_to_max(llp->pipe[0]); > + if (res > 0) > + llp->size = res; > + llp->can_grow = 0; > + goto fallback; > + } > + llp->size = res; > + } > + if (llp->size < pipesize) > + goto fallback; > + } > + > + > + res = vmsplice(llp->pipe[1], iov, iov_count, SPLICE_F_NONBLOCK); > + if (res == -1) > + goto fallback; > + > + if (res != headerlen) { > + res = -EIO; > + fuse_log(FUSE_LOG_ERR, "fuse: short vmsplice to pipe: > %u/%zu\n", res, > + headerlen); > + goto clear_pipe; > + } > + > + pipe_buf.buf[0].flags = FUSE_BUF_IS_FD; > + pipe_buf.buf[0].fd = llp->pipe[1]; > + > + res = fuse_buf_copy(&pipe_buf, buf, > + FUSE_BUF_FORCE_SPLICE | FUSE_BUF_SPLICE_NONBLOCK); > + if (res < 0) { > + if (res == -EAGAIN || res == -EINVAL) { > + /* > + * Should only get EAGAIN on kernels with > + * broken SPLICE_F_NONBLOCK support (<= > + * 2.6.35) where this error or a short read is > + * returned even if the pipe itself is not > + * full > + * > + * EINVAL might mean that splice can't handle > + * this combination of input and output. > + */ > + if (res == -EAGAIN) > + se->broken_splice_nonblock = 1; > + > + pthread_setspecific(se->pipe_key, NULL); > + fuse_ll_pipe_free(llp); > + goto fallback; > + } > + res = -res; > + goto clear_pipe; > + } > + > + if (res != 0 && res < len) { > + struct fuse_bufvec mem_buf = FUSE_BUFVEC_INIT(len); > + void *mbuf; > + size_t now_len = res; > + /* > + * For regular files a short count is either > + * 1) due to EOF, or > + * 2) because of broken SPLICE_F_NONBLOCK (see above) > + * > + * For other inputs it's possible that we overflowed > + * the pipe because of small buffer fragments. > + */ > + > + res = posix_memalign(&mbuf, pagesize, len); > + if (res != 0) > + goto clear_pipe; > + > + mem_buf.buf[0].mem = mbuf; > + mem_buf.off = now_len; > + res = fuse_buf_copy(&mem_buf, buf, 0); > + if (res > 0) { > + char *tmpbuf; > + size_t extra_len = res; > + /* > + * Trickiest case: got more data. Need to get > + * back the data from the pipe and then fall > + * back to regular write. > + */ > + tmpbuf = malloc(headerlen); > + if (tmpbuf == NULL) { > + free(mbuf); > + res = ENOMEM; > + goto clear_pipe; > + } > + res = read_back(llp->pipe[0], tmpbuf, headerlen); > + free(tmpbuf); > + if (res != 0) { > + free(mbuf); > + goto clear_pipe; > + } > + res = read_back(llp->pipe[0], mbuf, now_len); > + if (res != 0) { > + free(mbuf); > + goto clear_pipe; > + } > + len = now_len + extra_len; > + iov[iov_count].iov_base = mbuf; > + iov[iov_count].iov_len = len; > + iov_count++; > + res = fuse_send_msg(se, ch, iov, iov_count); > + free(mbuf); > + return res; > + } > + free(mbuf); > + res = now_len; > + } > + len = res; > + out->len = headerlen + len; > + > + if (se->debug) { > + fuse_log(FUSE_LOG_DEBUG, > + " unique: %llu, success, outsize: %i (splice)\n", > + (unsigned long long) out->unique, out->len); > + } > + > + splice_flags = 0; > + if ((flags & FUSE_BUF_SPLICE_MOVE) && > + (se->conn.want & FUSE_CAP_SPLICE_MOVE)) > + splice_flags |= SPLICE_F_MOVE; > + > + res = splice(llp->pipe[0], NULL, ch ? ch->fd : se->fd, > + NULL, out->len, splice_flags); > + if (res == -1) { > + res = -errno; > + perror("fuse: splice from pipe"); > + goto clear_pipe; > + } > + if (res != out->len) { > + res = -EIO; > + fuse_log(FUSE_LOG_ERR, "fuse: short splice from pipe: %u/%u\n", > + res, out->len); > + goto clear_pipe; > + } > + return 0; > + > +clear_pipe: > + fuse_ll_clear_pipe(se); > + return res; > + > +fallback: > + return fuse_send_data_iov_fallback(se, ch, iov, iov_count, buf, len); > +} > +#else > +static int fuse_send_data_iov(struct fuse_session *se, struct fuse_chan *ch, > + struct iovec *iov, int iov_count, > + struct fuse_bufvec *buf, unsigned int flags) > +{ > + size_t len = fuse_buf_size(buf); > + (void) flags; > + > + return fuse_send_data_iov_fallback(se, ch, iov, iov_count, buf, len); > +} > +#endif > + > +int fuse_reply_data(fuse_req_t req, struct fuse_bufvec *bufv, > + enum fuse_buf_copy_flags flags) > +{ > + struct iovec iov[2]; > + struct fuse_out_header out; > + int res; > + > + iov[0].iov_base = &out; > + iov[0].iov_len = sizeof(struct fuse_out_header); > + > + out.unique = req->unique; > + out.error = 0; > + > + res = fuse_send_data_iov(req->se, req->ch, iov, 1, bufv, flags); > + if (res <= 0) { > + fuse_free_req(req); > + return res; > + } else { > + return fuse_reply_err(req, res); > + } > +} > + > +int fuse_reply_statfs(fuse_req_t req, const struct statvfs *stbuf) > +{ > + struct fuse_statfs_out arg; > + size_t size = req->se->conn.proto_minor < 4 ? > + FUSE_COMPAT_STATFS_SIZE : sizeof(arg); > + > + memset(&arg, 0, sizeof(arg)); > + convert_statfs(stbuf, &arg.st); > + > + return send_reply_ok(req, &arg, size); > +} > + > +int fuse_reply_xattr(fuse_req_t req, size_t count) > +{ > + struct fuse_getxattr_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + arg.size = count; > + > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +int fuse_reply_lock(fuse_req_t req, const struct flock *lock) > +{ > + struct fuse_lk_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + arg.lk.type = lock->l_type; > + if (lock->l_type != F_UNLCK) { > + arg.lk.start = lock->l_start; > + if (lock->l_len == 0) > + arg.lk.end = OFFSET_MAX; > + else > + arg.lk.end = lock->l_start + lock->l_len - 1; > + } > + arg.lk.pid = lock->l_pid; > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +int fuse_reply_bmap(fuse_req_t req, uint64_t idx) > +{ > + struct fuse_bmap_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + arg.block = idx; > + > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +static struct fuse_ioctl_iovec *fuse_ioctl_iovec_copy(const struct iovec > *iov, > + size_t count) > +{ > + struct fuse_ioctl_iovec *fiov; > + size_t i; > + > + fiov = malloc(sizeof(fiov[0]) * count); > + if (!fiov) > + return NULL; > + > + for (i = 0; i < count; i++) { > + fiov[i].base = (uintptr_t) iov[i].iov_base; > + fiov[i].len = iov[i].iov_len; > + } > + > + return fiov; > +} > + > +int fuse_reply_ioctl_retry(fuse_req_t req, > + const struct iovec *in_iov, size_t in_count, > + const struct iovec *out_iov, size_t out_count) > +{ > + struct fuse_ioctl_out arg; > + struct fuse_ioctl_iovec *in_fiov = NULL; > + struct fuse_ioctl_iovec *out_fiov = NULL; > + struct iovec iov[4]; > + size_t count = 1; > + int res; > + > + memset(&arg, 0, sizeof(arg)); > + arg.flags |= FUSE_IOCTL_RETRY; > + arg.in_iovs = in_count; > + arg.out_iovs = out_count; > + iov[count].iov_base = &arg; > + iov[count].iov_len = sizeof(arg); > + count++; > + > + if (req->se->conn.proto_minor < 16) { > + if (in_count) { > + iov[count].iov_base = (void *)in_iov; > + iov[count].iov_len = sizeof(in_iov[0]) * in_count; > + count++; > + } > + > + if (out_count) { > + iov[count].iov_base = (void *)out_iov; > + iov[count].iov_len = sizeof(out_iov[0]) * out_count; > + count++; > + } > + } else { > + /* Can't handle non-compat 64bit ioctls on 32bit */ > + if (sizeof(void *) == 4 && req->ioctl_64bit) { > + res = fuse_reply_err(req, EINVAL); > + goto out; > + } > + > + if (in_count) { > + in_fiov = fuse_ioctl_iovec_copy(in_iov, in_count); > + if (!in_fiov) > + goto enomem; > + > + iov[count].iov_base = (void *)in_fiov; > + iov[count].iov_len = sizeof(in_fiov[0]) * in_count; > + count++; > + } > + if (out_count) { > + out_fiov = fuse_ioctl_iovec_copy(out_iov, out_count); > + if (!out_fiov) > + goto enomem; > + > + iov[count].iov_base = (void *)out_fiov; > + iov[count].iov_len = sizeof(out_fiov[0]) * out_count; > + count++; > + } > + } > + > + res = send_reply_iov(req, 0, iov, count); > +out: > + free(in_fiov); > + free(out_fiov); > + > + return res; > + > +enomem: > + res = fuse_reply_err(req, ENOMEM); > + goto out; > +} > + > +int fuse_reply_ioctl(fuse_req_t req, int result, const void *buf, size_t > size) > +{ > + struct fuse_ioctl_out arg; > + struct iovec iov[3]; > + size_t count = 1; > + > + memset(&arg, 0, sizeof(arg)); > + arg.result = result; > + iov[count].iov_base = &arg; > + iov[count].iov_len = sizeof(arg); > + count++; > + > + if (size) { > + iov[count].iov_base = (char *) buf; > + iov[count].iov_len = size; > + count++; > + } > + > + return send_reply_iov(req, 0, iov, count); > +} > + > +int fuse_reply_ioctl_iov(fuse_req_t req, int result, const struct iovec *iov, > + int count) > +{ > + struct iovec *padded_iov; > + struct fuse_ioctl_out arg; > + int res; > + > + padded_iov = malloc((count + 2) * sizeof(struct iovec)); > + if (padded_iov == NULL) > + return fuse_reply_err(req, ENOMEM); > + > + memset(&arg, 0, sizeof(arg)); > + arg.result = result; > + padded_iov[1].iov_base = &arg; > + padded_iov[1].iov_len = sizeof(arg); > + > + memcpy(&padded_iov[2], iov, count * sizeof(struct iovec)); > + > + res = send_reply_iov(req, 0, padded_iov, count + 2); > + free(padded_iov); > + > + return res; > +} > + > +int fuse_reply_poll(fuse_req_t req, unsigned revents) > +{ > + struct fuse_poll_out arg; > + > + memset(&arg, 0, sizeof(arg)); > + arg.revents = revents; > + > + return send_reply_ok(req, &arg, sizeof(arg)); > +} > + > +static void do_lookup(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + char *name = (char *) inarg; > + > + if (req->se->op.lookup) > + req->se->op.lookup(req, nodeid, name); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_forget(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_forget_in *arg = (struct fuse_forget_in *) inarg; > + > + if (req->se->op.forget) > + req->se->op.forget(req, nodeid, arg->nlookup); > + else > + fuse_reply_none(req); > +} > + > +static void do_batch_forget(fuse_req_t req, fuse_ino_t nodeid, > + const void *inarg) > +{ > + struct fuse_batch_forget_in *arg = (void *) inarg; > + struct fuse_forget_one *param = (void *) PARAM(arg); > + unsigned int i; > + > + (void) nodeid; > + > + if (req->se->op.forget_multi) { > + req->se->op.forget_multi(req, arg->count, > + (struct fuse_forget_data *) param); > + } else if (req->se->op.forget) { > + for (i = 0; i < arg->count; i++) { > + struct fuse_forget_one *forget = ¶m[i]; > + struct fuse_req *dummy_req; > + > + dummy_req = fuse_ll_alloc_req(req->se); > + if (dummy_req == NULL) > + break; > + > + dummy_req->unique = req->unique; > + dummy_req->ctx = req->ctx; > + dummy_req->ch = NULL; > + > + req->se->op.forget(dummy_req, forget->nodeid, > + forget->nlookup); > + } > + fuse_reply_none(req); > + } else { > + fuse_reply_none(req); > + } > +} > + > +static void do_getattr(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_file_info *fip = NULL; > + struct fuse_file_info fi; > + > + if (req->se->conn.proto_minor >= 9) { > + struct fuse_getattr_in *arg = (struct fuse_getattr_in *) inarg; > + > + if (arg->getattr_flags & FUSE_GETATTR_FH) { > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fip = &fi; > + } > + } > + > + if (req->se->op.getattr) > + req->se->op.getattr(req, nodeid, fip); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_setattr(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_setattr_in *arg = (struct fuse_setattr_in *) inarg; > + > + if (req->se->op.setattr) { > + struct fuse_file_info *fi = NULL; > + struct fuse_file_info fi_store; > + struct stat stbuf; > + memset(&stbuf, 0, sizeof(stbuf)); > + convert_attr(arg, &stbuf); > + if (arg->valid & FATTR_FH) { > + arg->valid &= ~FATTR_FH; > + memset(&fi_store, 0, sizeof(fi_store)); > + fi = &fi_store; > + fi->fh = arg->fh; > + } > + arg->valid &= > + FUSE_SET_ATTR_MODE | > + FUSE_SET_ATTR_UID | > + FUSE_SET_ATTR_GID | > + FUSE_SET_ATTR_SIZE | > + FUSE_SET_ATTR_ATIME | > + FUSE_SET_ATTR_MTIME | > + FUSE_SET_ATTR_ATIME_NOW | > + FUSE_SET_ATTR_MTIME_NOW | > + FUSE_SET_ATTR_CTIME; > + > + req->se->op.setattr(req, nodeid, &stbuf, arg->valid, fi); > + } else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_access(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_access_in *arg = (struct fuse_access_in *) inarg; > + > + if (req->se->op.access) > + req->se->op.access(req, nodeid, arg->mask); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_readlink(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + (void) inarg; > + > + if (req->se->op.readlink) > + req->se->op.readlink(req, nodeid); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_mknod(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_mknod_in *arg = (struct fuse_mknod_in *) inarg; > + char *name = PARAM(arg); > + > + if (req->se->conn.proto_minor >= 12) > + req->ctx.umask = arg->umask; > + else > + name = (char *) inarg + FUSE_COMPAT_MKNOD_IN_SIZE; > + > + if (req->se->op.mknod) > + req->se->op.mknod(req, nodeid, name, arg->mode, arg->rdev); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_mkdir(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_mkdir_in *arg = (struct fuse_mkdir_in *) inarg; > + > + if (req->se->conn.proto_minor >= 12) > + req->ctx.umask = arg->umask; > + > + if (req->se->op.mkdir) > + req->se->op.mkdir(req, nodeid, PARAM(arg), arg->mode); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_unlink(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + char *name = (char *) inarg; > + > + if (req->se->op.unlink) > + req->se->op.unlink(req, nodeid, name); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_rmdir(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + char *name = (char *) inarg; > + > + if (req->se->op.rmdir) > + req->se->op.rmdir(req, nodeid, name); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_symlink(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + char *name = (char *) inarg; > + char *linkname = ((char *) inarg) + strlen((char *) inarg) + 1; > + > + if (req->se->op.symlink) > + req->se->op.symlink(req, linkname, nodeid, name); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_rename(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_rename_in *arg = (struct fuse_rename_in *) inarg; > + char *oldname = PARAM(arg); > + char *newname = oldname + strlen(oldname) + 1; > + > + if (req->se->op.rename) > + req->se->op.rename(req, nodeid, oldname, arg->newdir, newname, > + 0); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_rename2(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_rename2_in *arg = (struct fuse_rename2_in *) inarg; > + char *oldname = PARAM(arg); > + char *newname = oldname + strlen(oldname) + 1; > + > + if (req->se->op.rename) > + req->se->op.rename(req, nodeid, oldname, arg->newdir, newname, > + arg->flags); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_link(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_link_in *arg = (struct fuse_link_in *) inarg; > + > + if (req->se->op.link) > + req->se->op.link(req, arg->oldnodeid, nodeid, PARAM(arg)); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_create(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_create_in *arg = (struct fuse_create_in *) inarg; > + > + if (req->se->op.create) { > + struct fuse_file_info fi; > + char *name = PARAM(arg); > + > + memset(&fi, 0, sizeof(fi)); > + fi.flags = arg->flags; > + > + if (req->se->conn.proto_minor >= 12) > + req->ctx.umask = arg->umask; > + else > + name = (char *) inarg + sizeof(struct fuse_open_in); > + > + req->se->op.create(req, nodeid, name, arg->mode, &fi); > + } else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_open(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_open_in *arg = (struct fuse_open_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.flags = arg->flags; > + > + if (req->se->op.open) > + req->se->op.open(req, nodeid, &fi); > + else > + fuse_reply_open(req, &fi); > +} > + > +static void do_read(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_read_in *arg = (struct fuse_read_in *) inarg; > + > + if (req->se->op.read) { > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + if (req->se->conn.proto_minor >= 9) { > + fi.lock_owner = arg->lock_owner; > + fi.flags = arg->flags; > + } > + req->se->op.read(req, nodeid, arg->size, arg->offset, &fi); > + } else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_write(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_write_in *arg = (struct fuse_write_in *) inarg; > + struct fuse_file_info fi; > + char *param; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.writepage = (arg->write_flags & FUSE_WRITE_CACHE) != 0; > + > + if (req->se->conn.proto_minor < 9) { > + param = ((char *) arg) + FUSE_COMPAT_WRITE_IN_SIZE; > + } else { > + fi.lock_owner = arg->lock_owner; > + fi.flags = arg->flags; > + param = PARAM(arg); > + } > + > + if (req->se->op.write) > + req->se->op.write(req, nodeid, param, arg->size, > + arg->offset, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_write_buf(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg, > + const struct fuse_buf *ibuf) > +{ > + struct fuse_session *se = req->se; > + struct fuse_bufvec bufv = { > + .buf[0] = *ibuf, > + .count = 1, > + }; > + struct fuse_write_in *arg = (struct fuse_write_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.writepage = arg->write_flags & FUSE_WRITE_CACHE; > + > + if (se->conn.proto_minor < 9) { > + bufv.buf[0].mem = ((char *) arg) + FUSE_COMPAT_WRITE_IN_SIZE; > + bufv.buf[0].size -= sizeof(struct fuse_in_header) + > + FUSE_COMPAT_WRITE_IN_SIZE; > + assert(!(bufv.buf[0].flags & FUSE_BUF_IS_FD)); > + } else { > + fi.lock_owner = arg->lock_owner; > + fi.flags = arg->flags; > + if (!(bufv.buf[0].flags & FUSE_BUF_IS_FD)) > + bufv.buf[0].mem = PARAM(arg); > + > + bufv.buf[0].size -= sizeof(struct fuse_in_header) + > + sizeof(struct fuse_write_in); > + } > + if (bufv.buf[0].size < arg->size) { > + fuse_log(FUSE_LOG_ERR, "fuse: do_write_buf: buffer size too > small\n"); > + fuse_reply_err(req, EIO); > + goto out; > + } > + bufv.buf[0].size = arg->size; > + > + se->op.write_buf(req, nodeid, &bufv, arg->offset, &fi); > + > +out: > + /* Need to reset the pipe if ->write_buf() didn't consume all data */ > + if ((ibuf->flags & FUSE_BUF_IS_FD) && bufv.idx < bufv.count) > + fuse_ll_clear_pipe(se); > +} > + > +static void do_flush(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_flush_in *arg = (struct fuse_flush_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.flush = 1; > + if (req->se->conn.proto_minor >= 7) > + fi.lock_owner = arg->lock_owner; > + > + if (req->se->op.flush) > + req->se->op.flush(req, nodeid, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_release(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_release_in *arg = (struct fuse_release_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.flags = arg->flags; > + fi.fh = arg->fh; > + if (req->se->conn.proto_minor >= 8) { > + fi.flush = (arg->release_flags & FUSE_RELEASE_FLUSH) ? 1 : 0; > + fi.lock_owner = arg->lock_owner; > + } > + if (arg->release_flags & FUSE_RELEASE_FLOCK_UNLOCK) { > + fi.flock_release = 1; > + fi.lock_owner = arg->lock_owner; > + } > + > + if (req->se->op.release) > + req->se->op.release(req, nodeid, &fi); > + else > + fuse_reply_err(req, 0); > +} > + > +static void do_fsync(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_fsync_in *arg = (struct fuse_fsync_in *) inarg; > + struct fuse_file_info fi; > + int datasync = arg->fsync_flags & 1; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (req->se->op.fsync) > + req->se->op.fsync(req, nodeid, datasync, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_opendir(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_open_in *arg = (struct fuse_open_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.flags = arg->flags; > + > + if (req->se->op.opendir) > + req->se->op.opendir(req, nodeid, &fi); > + else > + fuse_reply_open(req, &fi); > +} > + > +static void do_readdir(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_read_in *arg = (struct fuse_read_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (req->se->op.readdir) > + req->se->op.readdir(req, nodeid, arg->size, arg->offset, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_readdirplus(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + struct fuse_read_in *arg = (struct fuse_read_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (req->se->op.readdirplus) > + req->se->op.readdirplus(req, nodeid, arg->size, arg->offset, > &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_releasedir(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + struct fuse_release_in *arg = (struct fuse_release_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.flags = arg->flags; > + fi.fh = arg->fh; > + > + if (req->se->op.releasedir) > + req->se->op.releasedir(req, nodeid, &fi); > + else > + fuse_reply_err(req, 0); > +} > + > +static void do_fsyncdir(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_fsync_in *arg = (struct fuse_fsync_in *) inarg; > + struct fuse_file_info fi; > + int datasync = arg->fsync_flags & 1; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (req->se->op.fsyncdir) > + req->se->op.fsyncdir(req, nodeid, datasync, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_statfs(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + (void) nodeid; > + (void) inarg; > + > + if (req->se->op.statfs) > + req->se->op.statfs(req, nodeid); > + else { > + struct statvfs buf = { > + .f_namemax = 255, > + .f_bsize = 512, > + }; > + fuse_reply_statfs(req, &buf); > + } > +} > + > +static void do_setxattr(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_setxattr_in *arg = (struct fuse_setxattr_in *) inarg; > + char *name = PARAM(arg); > + char *value = name + strlen(name) + 1; > + > + if (req->se->op.setxattr) > + req->se->op.setxattr(req, nodeid, name, value, arg->size, > + arg->flags); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_getxattr(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_getxattr_in *arg = (struct fuse_getxattr_in *) inarg; > + > + if (req->se->op.getxattr) > + req->se->op.getxattr(req, nodeid, PARAM(arg), arg->size); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_listxattr(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + struct fuse_getxattr_in *arg = (struct fuse_getxattr_in *) inarg; > + > + if (req->se->op.listxattr) > + req->se->op.listxattr(req, nodeid, arg->size); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_removexattr(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + char *name = (char *) inarg; > + > + if (req->se->op.removexattr) > + req->se->op.removexattr(req, nodeid, name); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void convert_fuse_file_lock(struct fuse_file_lock *fl, > + struct flock *flock) > +{ > + memset(flock, 0, sizeof(struct flock)); > + flock->l_type = fl->type; > + flock->l_whence = SEEK_SET; > + flock->l_start = fl->start; > + if (fl->end == OFFSET_MAX) > + flock->l_len = 0; > + else > + flock->l_len = fl->end - fl->start + 1; > + flock->l_pid = fl->pid; > +} > + > +static void do_getlk(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_lk_in *arg = (struct fuse_lk_in *) inarg; > + struct fuse_file_info fi; > + struct flock flock; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.lock_owner = arg->owner; > + > + convert_fuse_file_lock(&arg->lk, &flock); > + if (req->se->op.getlk) > + req->se->op.getlk(req, nodeid, &fi, &flock); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_setlk_common(fuse_req_t req, fuse_ino_t nodeid, > + const void *inarg, int sleep) > +{ > + struct fuse_lk_in *arg = (struct fuse_lk_in *) inarg; > + struct fuse_file_info fi; > + struct flock flock; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.lock_owner = arg->owner; > + > + if (arg->lk_flags & FUSE_LK_FLOCK) { > + int op = 0; > + > + switch (arg->lk.type) { > + case F_RDLCK: > + op = LOCK_SH; > + break; > + case F_WRLCK: > + op = LOCK_EX; > + break; > + case F_UNLCK: > + op = LOCK_UN; > + break; > + } > + if (!sleep) > + op |= LOCK_NB; > + > + if (req->se->op.flock) > + req->se->op.flock(req, nodeid, &fi, op); > + else > + fuse_reply_err(req, ENOSYS); > + } else { > + convert_fuse_file_lock(&arg->lk, &flock); > + if (req->se->op.setlk) > + req->se->op.setlk(req, nodeid, &fi, &flock, sleep); > + else > + fuse_reply_err(req, ENOSYS); > + } > +} > + > +static void do_setlk(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + do_setlk_common(req, nodeid, inarg, 0); > +} > + > +static void do_setlkw(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + do_setlk_common(req, nodeid, inarg, 1); > +} > + > +static int find_interrupted(struct fuse_session *se, struct fuse_req *req) > +{ > + struct fuse_req *curr; > + > + for (curr = se->list.next; curr != &se->list; curr = curr->next) { > + if (curr->unique == req->u.i.unique) { > + fuse_interrupt_func_t func; > + void *data; > + > + curr->ctr++; > + pthread_mutex_unlock(&se->lock); > + > + /* Ugh, ugly locking */ > + pthread_mutex_lock(&curr->lock); > + pthread_mutex_lock(&se->lock); > + curr->interrupted = 1; > + func = curr->u.ni.func; > + data = curr->u.ni.data; > + pthread_mutex_unlock(&se->lock); > + if (func) > + func(curr, data); > + pthread_mutex_unlock(&curr->lock); > + > + pthread_mutex_lock(&se->lock); > + curr->ctr--; > + if (!curr->ctr) > + destroy_req(curr); > + > + return 1; > + } > + } > + for (curr = se->interrupts.next; curr != &se->interrupts; > + curr = curr->next) { > + if (curr->u.i.unique == req->u.i.unique) > + return 1; > + } > + return 0; > +} > + > +static void do_interrupt(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + struct fuse_interrupt_in *arg = (struct fuse_interrupt_in *) inarg; > + struct fuse_session *se = req->se; > + > + (void) nodeid; > + if (se->debug) > + fuse_log(FUSE_LOG_DEBUG, "INTERRUPT: %llu\n", > + (unsigned long long) arg->unique); > + > + req->u.i.unique = arg->unique; > + > + pthread_mutex_lock(&se->lock); > + if (find_interrupted(se, req)) > + destroy_req(req); > + else > + list_add_req(req, &se->interrupts); > + pthread_mutex_unlock(&se->lock); > +} > + > +static struct fuse_req *check_interrupt(struct fuse_session *se, > + struct fuse_req *req) > +{ > + struct fuse_req *curr; > + > + for (curr = se->interrupts.next; curr != &se->interrupts; > + curr = curr->next) { > + if (curr->u.i.unique == req->unique) { > + req->interrupted = 1; > + list_del_req(curr); > + free(curr); > + return NULL; > + } > + } > + curr = se->interrupts.next; > + if (curr != &se->interrupts) { > + list_del_req(curr); > + list_init_req(curr); > + return curr; > + } else > + return NULL; > +} > + > +static void do_bmap(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_bmap_in *arg = (struct fuse_bmap_in *) inarg; > + > + if (req->se->op.bmap) > + req->se->op.bmap(req, nodeid, arg->blocksize, arg->block); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_ioctl(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_ioctl_in *arg = (struct fuse_ioctl_in *) inarg; > + unsigned int flags = arg->flags; > + void *in_buf = arg->in_size ? PARAM(arg) : NULL; > + struct fuse_file_info fi; > + > + if (flags & FUSE_IOCTL_DIR && > + !(req->se->conn.want & FUSE_CAP_IOCTL_DIR)) { > + fuse_reply_err(req, ENOTTY); > + return; > + } > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (sizeof(void *) == 4 && req->se->conn.proto_minor >= 16 && > + !(flags & FUSE_IOCTL_32BIT)) { > + req->ioctl_64bit = 1; > + } > + > + if (req->se->op.ioctl) > + req->se->op.ioctl(req, nodeid, arg->cmd, > + (void *)(uintptr_t)arg->arg, &fi, flags, > + in_buf, arg->in_size, arg->out_size); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +void fuse_pollhandle_destroy(struct fuse_pollhandle *ph) > +{ > + free(ph); > +} > + > +static void do_poll(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_poll_in *arg = (struct fuse_poll_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + fi.poll_events = arg->events; > + > + if (req->se->op.poll) { > + struct fuse_pollhandle *ph = NULL; > + > + if (arg->flags & FUSE_POLL_SCHEDULE_NOTIFY) { > + ph = malloc(sizeof(struct fuse_pollhandle)); > + if (ph == NULL) { > + fuse_reply_err(req, ENOMEM); > + return; > + } > + ph->kh = arg->kh; > + ph->se = req->se; > + } > + > + req->se->op.poll(req, nodeid, &fi, ph); > + } else { > + fuse_reply_err(req, ENOSYS); > + } > +} > + > +static void do_fallocate(fuse_req_t req, fuse_ino_t nodeid, const void > *inarg) > +{ > + struct fuse_fallocate_in *arg = (struct fuse_fallocate_in *) inarg; > + struct fuse_file_info fi; > + > + memset(&fi, 0, sizeof(fi)); > + fi.fh = arg->fh; > + > + if (req->se->op.fallocate) > + req->se->op.fallocate(req, nodeid, arg->mode, arg->offset, > arg->length, &fi); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_copy_file_range(fuse_req_t req, fuse_ino_t nodeid_in, const > void *inarg) > +{ > + struct fuse_copy_file_range_in *arg = (struct fuse_copy_file_range_in > *) inarg; > + struct fuse_file_info fi_in, fi_out; > + > + memset(&fi_in, 0, sizeof(fi_in)); > + fi_in.fh = arg->fh_in; > + > + memset(&fi_out, 0, sizeof(fi_out)); > + fi_out.fh = arg->fh_out; > + > + > + if (req->se->op.copy_file_range) > + req->se->op.copy_file_range(req, nodeid_in, arg->off_in, > + &fi_in, arg->nodeid_out, > + arg->off_out, &fi_out, arg->len, > + arg->flags); > + else > + fuse_reply_err(req, ENOSYS); > +} > + > +static void do_init(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_init_in *arg = (struct fuse_init_in *) inarg; > + struct fuse_init_out outarg; > + struct fuse_session *se = req->se; > + size_t bufsize = se->bufsize; > + size_t outargsize = sizeof(outarg); > + > + (void) nodeid; > + if (se->debug) { > + fuse_log(FUSE_LOG_DEBUG, "INIT: %u.%u\n", arg->major, > arg->minor); > + if (arg->major == 7 && arg->minor >= 6) { > + fuse_log(FUSE_LOG_DEBUG, "flags=0x%08x\n", arg->flags); > + fuse_log(FUSE_LOG_DEBUG, "max_readahead=0x%08x\n", > + arg->max_readahead); > + } > + } > + se->conn.proto_major = arg->major; > + se->conn.proto_minor = arg->minor; > + se->conn.capable = 0; > + se->conn.want = 0; > + > + memset(&outarg, 0, sizeof(outarg)); > + outarg.major = FUSE_KERNEL_VERSION; > + outarg.minor = FUSE_KERNEL_MINOR_VERSION; > + > + if (arg->major < 7) { > + fuse_log(FUSE_LOG_ERR, "fuse: unsupported protocol version: > %u.%u\n", > + arg->major, arg->minor); > + fuse_reply_err(req, EPROTO); > + return; > + } > + > + if (arg->major > 7) { > + /* Wait for a second INIT request with a 7.X version */ > + send_reply_ok(req, &outarg, sizeof(outarg)); > + return; > + } > + > + if (arg->minor >= 6) { > + if (arg->max_readahead < se->conn.max_readahead) > + se->conn.max_readahead = arg->max_readahead; > + if (arg->flags & FUSE_ASYNC_READ) > + se->conn.capable |= FUSE_CAP_ASYNC_READ; > + if (arg->flags & FUSE_POSIX_LOCKS) > + se->conn.capable |= FUSE_CAP_POSIX_LOCKS; > + if (arg->flags & FUSE_ATOMIC_O_TRUNC) > + se->conn.capable |= FUSE_CAP_ATOMIC_O_TRUNC; > + if (arg->flags & FUSE_EXPORT_SUPPORT) > + se->conn.capable |= FUSE_CAP_EXPORT_SUPPORT; > + if (arg->flags & FUSE_DONT_MASK) > + se->conn.capable |= FUSE_CAP_DONT_MASK; > + if (arg->flags & FUSE_FLOCK_LOCKS) > + se->conn.capable |= FUSE_CAP_FLOCK_LOCKS; > + if (arg->flags & FUSE_AUTO_INVAL_DATA) > + se->conn.capable |= FUSE_CAP_AUTO_INVAL_DATA; > + if (arg->flags & FUSE_DO_READDIRPLUS) > + se->conn.capable |= FUSE_CAP_READDIRPLUS; > + if (arg->flags & FUSE_READDIRPLUS_AUTO) > + se->conn.capable |= FUSE_CAP_READDIRPLUS_AUTO; > + if (arg->flags & FUSE_ASYNC_DIO) > + se->conn.capable |= FUSE_CAP_ASYNC_DIO; > + if (arg->flags & FUSE_WRITEBACK_CACHE) > + se->conn.capable |= FUSE_CAP_WRITEBACK_CACHE; > + if (arg->flags & FUSE_NO_OPEN_SUPPORT) > + se->conn.capable |= FUSE_CAP_NO_OPEN_SUPPORT; > + if (arg->flags & FUSE_PARALLEL_DIROPS) > + se->conn.capable |= FUSE_CAP_PARALLEL_DIROPS; > + if (arg->flags & FUSE_POSIX_ACL) > + se->conn.capable |= FUSE_CAP_POSIX_ACL; > + if (arg->flags & FUSE_HANDLE_KILLPRIV) > + se->conn.capable |= FUSE_CAP_HANDLE_KILLPRIV; > + if (arg->flags & FUSE_NO_OPENDIR_SUPPORT) > + se->conn.capable |= FUSE_CAP_NO_OPENDIR_SUPPORT; > + if (!(arg->flags & FUSE_MAX_PAGES)) { > + size_t max_bufsize = > + FUSE_DEFAULT_MAX_PAGES_PER_REQ * getpagesize() > + + FUSE_BUFFER_HEADER_SIZE; > + if (bufsize > max_bufsize) { > + bufsize = max_bufsize; > + } > + } > + } else { > + se->conn.max_readahead = 0; > + } > + > + if (se->conn.proto_minor >= 14) { > +#ifdef HAVE_SPLICE > +#ifdef HAVE_VMSPLICE > + se->conn.capable |= FUSE_CAP_SPLICE_WRITE | > FUSE_CAP_SPLICE_MOVE; > +#endif > + se->conn.capable |= FUSE_CAP_SPLICE_READ; > +#endif > + } > + if (se->conn.proto_minor >= 18) > + se->conn.capable |= FUSE_CAP_IOCTL_DIR; > + > + /* Default settings for modern filesystems. > + * > + * Most of these capabilities were disabled by default in > + * libfuse2 for backwards compatibility reasons. In libfuse3, > + * we can finally enable them by default (as long as they're > + * supported by the kernel). > + */ > +#define LL_SET_DEFAULT(cond, cap) \ > + if ((cond) && (se->conn.capable & (cap))) \ > + se->conn.want |= (cap) > + LL_SET_DEFAULT(1, FUSE_CAP_ASYNC_READ); > + LL_SET_DEFAULT(1, FUSE_CAP_PARALLEL_DIROPS); > + LL_SET_DEFAULT(1, FUSE_CAP_AUTO_INVAL_DATA); > + LL_SET_DEFAULT(1, FUSE_CAP_HANDLE_KILLPRIV); > + LL_SET_DEFAULT(1, FUSE_CAP_ASYNC_DIO); > + LL_SET_DEFAULT(1, FUSE_CAP_IOCTL_DIR); > + LL_SET_DEFAULT(1, FUSE_CAP_ATOMIC_O_TRUNC); > + LL_SET_DEFAULT(se->op.write_buf, FUSE_CAP_SPLICE_READ); > + LL_SET_DEFAULT(se->op.getlk && se->op.setlk, > + FUSE_CAP_POSIX_LOCKS); > + LL_SET_DEFAULT(se->op.flock, FUSE_CAP_FLOCK_LOCKS); > + LL_SET_DEFAULT(se->op.readdirplus, FUSE_CAP_READDIRPLUS); > + LL_SET_DEFAULT(se->op.readdirplus && se->op.readdir, > + FUSE_CAP_READDIRPLUS_AUTO); > + se->conn.time_gran = 1; > + > + if (bufsize < FUSE_MIN_READ_BUFFER) { > + fuse_log(FUSE_LOG_ERR, "fuse: warning: buffer size too small: > %zu\n", > + bufsize); > + bufsize = FUSE_MIN_READ_BUFFER; > + } > + se->bufsize = bufsize; > + > + if (se->conn.max_write > bufsize - FUSE_BUFFER_HEADER_SIZE) > + se->conn.max_write = bufsize - FUSE_BUFFER_HEADER_SIZE; > + > + se->got_init = 1; > + if (se->op.init) > + se->op.init(se->userdata, &se->conn); > + > + if (se->conn.want & (~se->conn.capable)) { > + fuse_log(FUSE_LOG_ERR, "fuse: error: filesystem requested > capabilities " > + "0x%x that are not supported by kernel, aborting.\n", > + se->conn.want & (~se->conn.capable)); > + fuse_reply_err(req, EPROTO); > + se->error = -EPROTO; > + fuse_session_exit(se); > + return; > + } > + > + unsigned max_read_mo = get_max_read(se->mo); > + if (se->conn.max_read != max_read_mo) { > + fuse_log(FUSE_LOG_ERR, "fuse: error: init() and > fuse_session_new() " > + "requested different maximum read size (%u vs %u)\n", > + se->conn.max_read, max_read_mo); > + fuse_reply_err(req, EPROTO); > + se->error = -EPROTO; > + fuse_session_exit(se); > + return; > + } > + > + if (se->conn.max_write < bufsize - FUSE_BUFFER_HEADER_SIZE) { > + se->bufsize = se->conn.max_write + FUSE_BUFFER_HEADER_SIZE; > + } > + if (arg->flags & FUSE_MAX_PAGES) { > + outarg.flags |= FUSE_MAX_PAGES; > + outarg.max_pages = (se->conn.max_write - 1) / getpagesize() + 1; > + } > + > + /* Always enable big writes, this is superseded > + by the max_write option */ > + outarg.flags |= FUSE_BIG_WRITES; > + > + if (se->conn.want & FUSE_CAP_ASYNC_READ) > + outarg.flags |= FUSE_ASYNC_READ; > + if (se->conn.want & FUSE_CAP_POSIX_LOCKS) > + outarg.flags |= FUSE_POSIX_LOCKS; > + if (se->conn.want & FUSE_CAP_ATOMIC_O_TRUNC) > + outarg.flags |= FUSE_ATOMIC_O_TRUNC; > + if (se->conn.want & FUSE_CAP_EXPORT_SUPPORT) > + outarg.flags |= FUSE_EXPORT_SUPPORT; > + if (se->conn.want & FUSE_CAP_DONT_MASK) > + outarg.flags |= FUSE_DONT_MASK; > + if (se->conn.want & FUSE_CAP_FLOCK_LOCKS) > + outarg.flags |= FUSE_FLOCK_LOCKS; > + if (se->conn.want & FUSE_CAP_AUTO_INVAL_DATA) > + outarg.flags |= FUSE_AUTO_INVAL_DATA; > + if (se->conn.want & FUSE_CAP_READDIRPLUS) > + outarg.flags |= FUSE_DO_READDIRPLUS; > + if (se->conn.want & FUSE_CAP_READDIRPLUS_AUTO) > + outarg.flags |= FUSE_READDIRPLUS_AUTO; > + if (se->conn.want & FUSE_CAP_ASYNC_DIO) > + outarg.flags |= FUSE_ASYNC_DIO; > + if (se->conn.want & FUSE_CAP_WRITEBACK_CACHE) > + outarg.flags |= FUSE_WRITEBACK_CACHE; > + if (se->conn.want & FUSE_CAP_POSIX_ACL) > + outarg.flags |= FUSE_POSIX_ACL; > + outarg.max_readahead = se->conn.max_readahead; > + outarg.max_write = se->conn.max_write; > + if (se->conn.proto_minor >= 13) { > + if (se->conn.max_background >= (1 << 16)) > + se->conn.max_background = (1 << 16) - 1; > + if (se->conn.congestion_threshold > se->conn.max_background) > + se->conn.congestion_threshold = se->conn.max_background; > + if (!se->conn.congestion_threshold) { > + se->conn.congestion_threshold = > + se->conn.max_background * 3 / 4; > + } > + > + outarg.max_background = se->conn.max_background; > + outarg.congestion_threshold = se->conn.congestion_threshold; > + } > + if (se->conn.proto_minor >= 23) > + outarg.time_gran = se->conn.time_gran; > + > + if (se->debug) { > + fuse_log(FUSE_LOG_DEBUG, " INIT: %u.%u\n", outarg.major, > outarg.minor); > + fuse_log(FUSE_LOG_DEBUG, " flags=0x%08x\n", outarg.flags); > + fuse_log(FUSE_LOG_DEBUG, " max_readahead=0x%08x\n", > + outarg.max_readahead); > + fuse_log(FUSE_LOG_DEBUG, " max_write=0x%08x\n", > outarg.max_write); > + fuse_log(FUSE_LOG_DEBUG, " max_background=%i\n", > + outarg.max_background); > + fuse_log(FUSE_LOG_DEBUG, " congestion_threshold=%i\n", > + outarg.congestion_threshold); > + fuse_log(FUSE_LOG_DEBUG, " time_gran=%u\n", > + outarg.time_gran); > + } > + if (arg->minor < 5) > + outargsize = FUSE_COMPAT_INIT_OUT_SIZE; > + else if (arg->minor < 23) > + outargsize = FUSE_COMPAT_22_INIT_OUT_SIZE; > + > + send_reply_ok(req, &outarg, outargsize); > +} > + > +static void do_destroy(fuse_req_t req, fuse_ino_t nodeid, const void *inarg) > +{ > + struct fuse_session *se = req->se; > + > + (void) nodeid; > + (void) inarg; > + > + se->got_destroy = 1; > + if (se->op.destroy) > + se->op.destroy(se->userdata); > + > + send_reply_ok(req, NULL, 0); > +} > + > +static void list_del_nreq(struct fuse_notify_req *nreq) > +{ > + struct fuse_notify_req *prev = nreq->prev; > + struct fuse_notify_req *next = nreq->next; > + prev->next = next; > + next->prev = prev; > +} > + > +static void list_add_nreq(struct fuse_notify_req *nreq, > + struct fuse_notify_req *next) > +{ > + struct fuse_notify_req *prev = next->prev; > + nreq->next = next; > + nreq->prev = prev; > + prev->next = nreq; > + next->prev = nreq; > +} > + > +static void list_init_nreq(struct fuse_notify_req *nreq) > +{ > + nreq->next = nreq; > + nreq->prev = nreq; > +} > + > +static void do_notify_reply(fuse_req_t req, fuse_ino_t nodeid, > + const void *inarg, const struct fuse_buf *buf) > +{ > + struct fuse_session *se = req->se; > + struct fuse_notify_req *nreq; > + struct fuse_notify_req *head; > + > + pthread_mutex_lock(&se->lock); > + head = &se->notify_list; > + for (nreq = head->next; nreq != head; nreq = nreq->next) { > + if (nreq->unique == req->unique) { > + list_del_nreq(nreq); > + break; > + } > + } > + pthread_mutex_unlock(&se->lock); > + > + if (nreq != head) > + nreq->reply(nreq, req, nodeid, inarg, buf); > +} > + > +static int send_notify_iov(struct fuse_session *se, int notify_code, > + struct iovec *iov, int count) > +{ > + struct fuse_out_header out; > + > + if (!se->got_init) > + return -ENOTCONN; > + > + out.unique = 0; > + out.error = notify_code; > + iov[0].iov_base = &out; > + iov[0].iov_len = sizeof(struct fuse_out_header); > + > + return fuse_send_msg(se, NULL, iov, count); > +} > + > +int fuse_lowlevel_notify_poll(struct fuse_pollhandle *ph) > +{ > + if (ph != NULL) { > + struct fuse_notify_poll_wakeup_out outarg; > + struct iovec iov[2]; > + > + outarg.kh = ph->kh; > + > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + > + return send_notify_iov(ph->se, FUSE_NOTIFY_POLL, iov, 2); > + } else { > + return 0; > + } > +} > + > +int fuse_lowlevel_notify_inval_inode(struct fuse_session *se, fuse_ino_t ino, > + off_t off, off_t len) > +{ > + struct fuse_notify_inval_inode_out outarg; > + struct iovec iov[2]; > + > + if (!se) > + return -EINVAL; > + > + if (se->conn.proto_major < 6 || se->conn.proto_minor < 12) > + return -ENOSYS; > + > + outarg.ino = ino; > + outarg.off = off; > + outarg.len = len; > + > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + > + return send_notify_iov(se, FUSE_NOTIFY_INVAL_INODE, iov, 2); > +} > + > +int fuse_lowlevel_notify_inval_entry(struct fuse_session *se, fuse_ino_t > parent, > + const char *name, size_t namelen) > +{ > + struct fuse_notify_inval_entry_out outarg; > + struct iovec iov[3]; > + > + if (!se) > + return -EINVAL; > + > + if (se->conn.proto_major < 6 || se->conn.proto_minor < 12) > + return -ENOSYS; > + > + outarg.parent = parent; > + outarg.namelen = namelen; > + outarg.padding = 0; > + > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + iov[2].iov_base = (void *)name; > + iov[2].iov_len = namelen + 1; > + > + return send_notify_iov(se, FUSE_NOTIFY_INVAL_ENTRY, iov, 3); > +} > + > +int fuse_lowlevel_notify_delete(struct fuse_session *se, > + fuse_ino_t parent, fuse_ino_t child, > + const char *name, size_t namelen) > +{ > + struct fuse_notify_delete_out outarg; > + struct iovec iov[3]; > + > + if (!se) > + return -EINVAL; > + > + if (se->conn.proto_major < 6 || se->conn.proto_minor < 18) > + return -ENOSYS; > + > + outarg.parent = parent; > + outarg.child = child; > + outarg.namelen = namelen; > + outarg.padding = 0; > + > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + iov[2].iov_base = (void *)name; > + iov[2].iov_len = namelen + 1; > + > + return send_notify_iov(se, FUSE_NOTIFY_DELETE, iov, 3); > +} > + > +int fuse_lowlevel_notify_store(struct fuse_session *se, fuse_ino_t ino, > + off_t offset, struct fuse_bufvec *bufv, > + enum fuse_buf_copy_flags flags) > +{ > + struct fuse_out_header out; > + struct fuse_notify_store_out outarg; > + struct iovec iov[3]; > + size_t size = fuse_buf_size(bufv); > + int res; > + > + if (!se) > + return -EINVAL; > + > + if (se->conn.proto_major < 6 || se->conn.proto_minor < 15) > + return -ENOSYS; > + > + out.unique = 0; > + out.error = FUSE_NOTIFY_STORE; > + > + outarg.nodeid = ino; > + outarg.offset = offset; > + outarg.size = size; > + outarg.padding = 0; > + > + iov[0].iov_base = &out; > + iov[0].iov_len = sizeof(out); > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + > + res = fuse_send_data_iov(se, NULL, iov, 2, bufv, flags); > + if (res > 0) > + res = -res; > + > + return res; > +} > + > +struct fuse_retrieve_req { > + struct fuse_notify_req nreq; > + void *cookie; > +}; > + > +static void fuse_ll_retrieve_reply(struct fuse_notify_req *nreq, > + fuse_req_t req, fuse_ino_t ino, > + const void *inarg, > + const struct fuse_buf *ibuf) > +{ > + struct fuse_session *se = req->se; > + struct fuse_retrieve_req *rreq = > + container_of(nreq, struct fuse_retrieve_req, nreq); > + const struct fuse_notify_retrieve_in *arg = inarg; > + struct fuse_bufvec bufv = { > + .buf[0] = *ibuf, > + .count = 1, > + }; > + > + if (!(bufv.buf[0].flags & FUSE_BUF_IS_FD)) > + bufv.buf[0].mem = PARAM(arg); > + > + bufv.buf[0].size -= sizeof(struct fuse_in_header) + > + sizeof(struct fuse_notify_retrieve_in); > + > + if (bufv.buf[0].size < arg->size) { > + fuse_log(FUSE_LOG_ERR, "fuse: retrieve reply: buffer size too > small\n"); > + fuse_reply_none(req); > + goto out; > + } > + bufv.buf[0].size = arg->size; > + > + if (se->op.retrieve_reply) { > + se->op.retrieve_reply(req, rreq->cookie, ino, > + arg->offset, &bufv); > + } else { > + fuse_reply_none(req); > + } > +out: > + free(rreq); > + if ((ibuf->flags & FUSE_BUF_IS_FD) && bufv.idx < bufv.count) > + fuse_ll_clear_pipe(se); > +} > + > +int fuse_lowlevel_notify_retrieve(struct fuse_session *se, fuse_ino_t ino, > + size_t size, off_t offset, void *cookie) > +{ > + struct fuse_notify_retrieve_out outarg; > + struct iovec iov[2]; > + struct fuse_retrieve_req *rreq; > + int err; > + > + if (!se) > + return -EINVAL; > + > + if (se->conn.proto_major < 6 || se->conn.proto_minor < 15) > + return -ENOSYS; > + > + rreq = malloc(sizeof(*rreq)); > + if (rreq == NULL) > + return -ENOMEM; > + > + pthread_mutex_lock(&se->lock); > + rreq->cookie = cookie; > + rreq->nreq.unique = se->notify_ctr++; > + rreq->nreq.reply = fuse_ll_retrieve_reply; > + list_add_nreq(&rreq->nreq, &se->notify_list); > + pthread_mutex_unlock(&se->lock); > + > + outarg.notify_unique = rreq->nreq.unique; > + outarg.nodeid = ino; > + outarg.offset = offset; > + outarg.size = size; > + outarg.padding = 0; > + > + iov[1].iov_base = &outarg; > + iov[1].iov_len = sizeof(outarg); > + > + err = send_notify_iov(se, FUSE_NOTIFY_RETRIEVE, iov, 2); > + if (err) { > + pthread_mutex_lock(&se->lock); > + list_del_nreq(&rreq->nreq); > + pthread_mutex_unlock(&se->lock); > + free(rreq); > + } > + > + return err; > +} > + > +void *fuse_req_userdata(fuse_req_t req) > +{ > + return req->se->userdata; > +} > + > +const struct fuse_ctx *fuse_req_ctx(fuse_req_t req) > +{ > + return &req->ctx; > +} > + > +void fuse_req_interrupt_func(fuse_req_t req, fuse_interrupt_func_t func, > + void *data) > +{ > + pthread_mutex_lock(&req->lock); > + pthread_mutex_lock(&req->se->lock); > + req->u.ni.func = func; > + req->u.ni.data = data; > + pthread_mutex_unlock(&req->se->lock); > + if (req->interrupted && func) > + func(req, data); > + pthread_mutex_unlock(&req->lock); > +} > + > +int fuse_req_interrupted(fuse_req_t req) > +{ > + int interrupted; > + > + pthread_mutex_lock(&req->se->lock); > + interrupted = req->interrupted; > + pthread_mutex_unlock(&req->se->lock); > + > + return interrupted; > +} > + > +static struct { > + void (*func)(fuse_req_t, fuse_ino_t, const void *); > + const char *name; > +} fuse_ll_ops[] = { > + [FUSE_LOOKUP] = { do_lookup, "LOOKUP" }, > + [FUSE_FORGET] = { do_forget, "FORGET" }, > + [FUSE_GETATTR] = { do_getattr, "GETATTR" }, > + [FUSE_SETATTR] = { do_setattr, "SETATTR" }, > + [FUSE_READLINK] = { do_readlink, "READLINK" }, > + [FUSE_SYMLINK] = { do_symlink, "SYMLINK" }, > + [FUSE_MKNOD] = { do_mknod, "MKNOD" }, > + [FUSE_MKDIR] = { do_mkdir, "MKDIR" }, > + [FUSE_UNLINK] = { do_unlink, "UNLINK" }, > + [FUSE_RMDIR] = { do_rmdir, "RMDIR" }, > + [FUSE_RENAME] = { do_rename, "RENAME" }, > + [FUSE_LINK] = { do_link, "LINK" }, > + [FUSE_OPEN] = { do_open, "OPEN" }, > + [FUSE_READ] = { do_read, "READ" }, > + [FUSE_WRITE] = { do_write, "WRITE" }, > + [FUSE_STATFS] = { do_statfs, "STATFS" }, > + [FUSE_RELEASE] = { do_release, "RELEASE" }, > + [FUSE_FSYNC] = { do_fsync, "FSYNC" }, > + [FUSE_SETXATTR] = { do_setxattr, "SETXATTR" }, > + [FUSE_GETXATTR] = { do_getxattr, "GETXATTR" }, > + [FUSE_LISTXATTR] = { do_listxattr, "LISTXATTR" }, > + [FUSE_REMOVEXATTR] = { do_removexattr, "REMOVEXATTR" }, > + [FUSE_FLUSH] = { do_flush, "FLUSH" }, > + [FUSE_INIT] = { do_init, "INIT" }, > + [FUSE_OPENDIR] = { do_opendir, "OPENDIR" }, > + [FUSE_READDIR] = { do_readdir, "READDIR" }, > + [FUSE_RELEASEDIR] = { do_releasedir, "RELEASEDIR" }, > + [FUSE_FSYNCDIR] = { do_fsyncdir, "FSYNCDIR" }, > + [FUSE_GETLK] = { do_getlk, "GETLK" }, > + [FUSE_SETLK] = { do_setlk, "SETLK" }, > + [FUSE_SETLKW] = { do_setlkw, "SETLKW" }, > + [FUSE_ACCESS] = { do_access, "ACCESS" }, > + [FUSE_CREATE] = { do_create, "CREATE" }, > + [FUSE_INTERRUPT] = { do_interrupt, "INTERRUPT" }, > + [FUSE_BMAP] = { do_bmap, "BMAP" }, > + [FUSE_IOCTL] = { do_ioctl, "IOCTL" }, > + [FUSE_POLL] = { do_poll, "POLL" }, > + [FUSE_FALLOCATE] = { do_fallocate, "FALLOCATE" }, > + [FUSE_DESTROY] = { do_destroy, "DESTROY" }, > + [FUSE_NOTIFY_REPLY] = { (void *) 1, "NOTIFY_REPLY" }, > + [FUSE_BATCH_FORGET] = { do_batch_forget, "BATCH_FORGET" }, > + [FUSE_READDIRPLUS] = { do_readdirplus, "READDIRPLUS"}, > + [FUSE_RENAME2] = { do_rename2, "RENAME2" }, > + [FUSE_COPY_FILE_RANGE] = { do_copy_file_range, "COPY_FILE_RANGE" }, > + [CUSE_INIT] = { cuse_lowlevel_init, "CUSE_INIT" }, > +}; > + > +#define FUSE_MAXOP (sizeof(fuse_ll_ops) / sizeof(fuse_ll_ops[0])) > + > +static const char *opname(enum fuse_opcode opcode) > +{ > + if (opcode >= FUSE_MAXOP || !fuse_ll_ops[opcode].name) > + return "???"; > + else > + return fuse_ll_ops[opcode].name; > +} > + > +static int fuse_ll_copy_from_pipe(struct fuse_bufvec *dst, > + struct fuse_bufvec *src) > +{ > + ssize_t res = fuse_buf_copy(dst, src, 0); > + if (res < 0) { > + fuse_log(FUSE_LOG_ERR, "fuse: copy from pipe: %s\n", > strerror(-res)); > + return res; > + } > + if ((size_t)res < fuse_buf_size(dst)) { > + fuse_log(FUSE_LOG_ERR, "fuse: copy from pipe: short read\n"); > + return -1; > + } > + return 0; > +} > + > +void fuse_session_process_buf(struct fuse_session *se, > + const struct fuse_buf *buf) > +{ > + fuse_session_process_buf_int(se, buf, NULL); > +} > + > +void fuse_session_process_buf_int(struct fuse_session *se, > + const struct fuse_buf *buf, struct fuse_chan > *ch) > +{ > + const size_t write_header_size = sizeof(struct fuse_in_header) + > + sizeof(struct fuse_write_in); > + struct fuse_bufvec bufv = { .buf[0] = *buf, .count = 1 }; > + struct fuse_bufvec tmpbuf = FUSE_BUFVEC_INIT(write_header_size); > + struct fuse_in_header *in; > + const void *inarg; > + struct fuse_req *req; > + void *mbuf = NULL; > + int err; > + int res; > + > + if (buf->flags & FUSE_BUF_IS_FD) { > + if (buf->size < tmpbuf.buf[0].size) > + tmpbuf.buf[0].size = buf->size; > + > + mbuf = malloc(tmpbuf.buf[0].size); > + if (mbuf == NULL) { > + fuse_log(FUSE_LOG_ERR, "fuse: failed to allocate > header\n"); > + goto clear_pipe; > + } > + tmpbuf.buf[0].mem = mbuf; > + > + res = fuse_ll_copy_from_pipe(&tmpbuf, &bufv); > + if (res < 0) > + goto clear_pipe; > + > + in = mbuf; > + } else { > + in = buf->mem; > + } > + > + if (se->debug) { > + fuse_log(FUSE_LOG_DEBUG, > + "unique: %llu, opcode: %s (%i), nodeid: %llu, insize: > %zu, pid: %u\n", > + (unsigned long long) in->unique, > + opname((enum fuse_opcode) in->opcode), in->opcode, > + (unsigned long long) in->nodeid, buf->size, in->pid); > + } > + > + req = fuse_ll_alloc_req(se); > + if (req == NULL) { > + struct fuse_out_header out = { > + .unique = in->unique, > + .error = -ENOMEM, > + }; > + struct iovec iov = { > + .iov_base = &out, > + .iov_len = sizeof(struct fuse_out_header), > + }; > + > + fuse_send_msg(se, ch, &iov, 1); > + goto clear_pipe; > + } > + > + req->unique = in->unique; > + req->ctx.uid = in->uid; > + req->ctx.gid = in->gid; > + req->ctx.pid = in->pid; > + req->ch = ch ? fuse_chan_get(ch) : NULL; > + > + err = EIO; > + if (!se->got_init) { > + enum fuse_opcode expected; > + > + expected = se->cuse_data ? CUSE_INIT : FUSE_INIT; > + if (in->opcode != expected) > + goto reply_err; > + } else if (in->opcode == FUSE_INIT || in->opcode == CUSE_INIT) > + goto reply_err; > + > + err = EACCES; > + /* Implement -o allow_root */ > + if (se->deny_others && in->uid != se->owner && in->uid != 0 && > + in->opcode != FUSE_INIT && in->opcode != FUSE_READ && > + in->opcode != FUSE_WRITE && in->opcode != FUSE_FSYNC && > + in->opcode != FUSE_RELEASE && in->opcode != FUSE_READDIR && > + in->opcode != FUSE_FSYNCDIR && in->opcode != FUSE_RELEASEDIR && > + in->opcode != FUSE_NOTIFY_REPLY && > + in->opcode != FUSE_READDIRPLUS) > + goto reply_err; > + > + err = ENOSYS; > + if (in->opcode >= FUSE_MAXOP || !fuse_ll_ops[in->opcode].func) > + goto reply_err; > + if (in->opcode != FUSE_INTERRUPT) { > + struct fuse_req *intr; > + pthread_mutex_lock(&se->lock); > + intr = check_interrupt(se, req); > + list_add_req(req, &se->list); > + pthread_mutex_unlock(&se->lock); > + if (intr) > + fuse_reply_err(intr, EAGAIN); > + } > + > + if ((buf->flags & FUSE_BUF_IS_FD) && write_header_size < buf->size && > + (in->opcode != FUSE_WRITE || !se->op.write_buf) && > + in->opcode != FUSE_NOTIFY_REPLY) { > + void *newmbuf; > + > + err = ENOMEM; > + newmbuf = realloc(mbuf, buf->size); > + if (newmbuf == NULL) > + goto reply_err; > + mbuf = newmbuf; > + > + tmpbuf = FUSE_BUFVEC_INIT(buf->size - write_header_size); > + tmpbuf.buf[0].mem = (char *)mbuf + write_header_size; > + > + res = fuse_ll_copy_from_pipe(&tmpbuf, &bufv); > + err = -res; > + if (res < 0) > + goto reply_err; > + > + in = mbuf; > + } > + > + inarg = (void *) &in[1]; > + if (in->opcode == FUSE_WRITE && se->op.write_buf) > + do_write_buf(req, in->nodeid, inarg, buf); > + else if (in->opcode == FUSE_NOTIFY_REPLY) > + do_notify_reply(req, in->nodeid, inarg, buf); > + else > + fuse_ll_ops[in->opcode].func(req, in->nodeid, inarg); > + > +out_free: > + free(mbuf); > + return; > + > +reply_err: > + fuse_reply_err(req, err); > +clear_pipe: > + if (buf->flags & FUSE_BUF_IS_FD) > + fuse_ll_clear_pipe(se); > + goto out_free; > +} > + > +#define LL_OPTION(n,o,v) \ > + { n, offsetof(struct fuse_session, o), v } > + > +static const struct fuse_opt fuse_ll_opts[] = { > + LL_OPTION("debug", debug, 1), > + LL_OPTION("-d", debug, 1), > + LL_OPTION("--debug", debug, 1), > + LL_OPTION("allow_root", deny_others, 1), > + FUSE_OPT_END > +}; > + > +void fuse_lowlevel_version(void) > +{ > + printf("using FUSE kernel interface version %i.%i\n", > + FUSE_KERNEL_VERSION, FUSE_KERNEL_MINOR_VERSION); > + fuse_mount_version(); > +} > + > +void fuse_lowlevel_help(void) > +{ > + /* These are not all options, but the ones that are > + potentially of interest to an end-user */ > + printf( > +" -o allow_other allow access by all users\n" > +" -o allow_root allow access by root\n" > +" -o auto_unmount auto unmount on process termination\n"); > +} > + > +void fuse_session_destroy(struct fuse_session *se) > +{ > + struct fuse_ll_pipe *llp; > + > + if (se->got_init && !se->got_destroy) { > + if (se->op.destroy) > + se->op.destroy(se->userdata); > + } > + llp = pthread_getspecific(se->pipe_key); > + if (llp != NULL) > + fuse_ll_pipe_free(llp); > + pthread_key_delete(se->pipe_key); > + pthread_mutex_destroy(&se->lock); > + free(se->cuse_data); > + if (se->fd != -1) > + close(se->fd); > + destroy_mount_opts(se->mo); > + free(se); > +} > + > + > +static void fuse_ll_pipe_destructor(void *data) > +{ > + struct fuse_ll_pipe *llp = data; > + fuse_ll_pipe_free(llp); > +} > + > +int fuse_session_receive_buf(struct fuse_session *se, struct fuse_buf *buf) > +{ > + return fuse_session_receive_buf_int(se, buf, NULL); > +} > + > +int fuse_session_receive_buf_int(struct fuse_session *se, struct fuse_buf > *buf, > + struct fuse_chan *ch) > +{ > + int err; > + ssize_t res; > +#ifdef HAVE_SPLICE > + size_t bufsize = se->bufsize; > + struct fuse_ll_pipe *llp; > + struct fuse_buf tmpbuf; > + > + if (se->conn.proto_minor < 14 || !(se->conn.want & > FUSE_CAP_SPLICE_READ)) > + goto fallback; > + > + llp = fuse_ll_get_pipe(se); > + if (llp == NULL) > + goto fallback; > + > + if (llp->size < bufsize) { > + if (llp->can_grow) { > + res = fcntl(llp->pipe[0], F_SETPIPE_SZ, bufsize); > + if (res == -1) { > + llp->can_grow = 0; > + res = grow_pipe_to_max(llp->pipe[0]); > + if (res > 0) > + llp->size = res; > + goto fallback; > + } > + llp->size = res; > + } > + if (llp->size < bufsize) > + goto fallback; > + } > + > + res = splice(ch ? ch->fd : se->fd, > + NULL, llp->pipe[1], NULL, bufsize, 0); > + err = errno; > + > + if (fuse_session_exited(se)) > + return 0; > + > + if (res == -1) { > + if (err == ENODEV) { > + /* Filesystem was unmounted, or connection was aborted > + via /sys/fs/fuse/connections */ > + fuse_session_exit(se); > + return 0; > + } > + if (err != EINTR && err != EAGAIN) > + perror("fuse: splice from device"); > + return -err; > + } > + > + if (res < sizeof(struct fuse_in_header)) { > + fuse_log(FUSE_LOG_ERR, "short splice from fuse device\n"); > + return -EIO; > + } > + > + tmpbuf = (struct fuse_buf) { > + .size = res, > + .flags = FUSE_BUF_IS_FD, > + .fd = llp->pipe[0], > + }; > + > + /* > + * Don't bother with zero copy for small requests. > + * fuse_loop_mt() needs to check for FORGET so this more than > + * just an optimization. > + */ > + if (res < sizeof(struct fuse_in_header) + > + sizeof(struct fuse_write_in) + pagesize) { > + struct fuse_bufvec src = { .buf[0] = tmpbuf, .count = 1 }; > + struct fuse_bufvec dst = { .count = 1 }; > + > + if (!buf->mem) { > + buf->mem = malloc(se->bufsize); > + if (!buf->mem) { > + fuse_log(FUSE_LOG_ERR, > + "fuse: failed to allocate read > buffer\n"); > + return -ENOMEM; > + } > + } > + buf->size = se->bufsize; > + buf->flags = 0; > + dst.buf[0] = *buf; > + > + res = fuse_buf_copy(&dst, &src, 0); > + if (res < 0) { > + fuse_log(FUSE_LOG_ERR, "fuse: copy from pipe: %s\n", > + strerror(-res)); > + fuse_ll_clear_pipe(se); > + return res; > + } > + if (res < tmpbuf.size) { > + fuse_log(FUSE_LOG_ERR, "fuse: copy from pipe: short > read\n"); > + fuse_ll_clear_pipe(se); > + return -EIO; > + } > + assert(res == tmpbuf.size); > + > + } else { > + /* Don't overwrite buf->mem, as that would cause a leak */ > + buf->fd = tmpbuf.fd; > + buf->flags = tmpbuf.flags; > + } > + buf->size = tmpbuf.size; > + > + return res; > + > +fallback: > +#endif > + if (!buf->mem) { > + buf->mem = malloc(se->bufsize); > + if (!buf->mem) { > + fuse_log(FUSE_LOG_ERR, > + "fuse: failed to allocate read buffer\n"); > + return -ENOMEM; > + } > + } > + > +restart: > + res = read(ch ? ch->fd : se->fd, buf->mem, se->bufsize); > + err = errno; > + > + if (fuse_session_exited(se)) > + return 0; > + if (res == -1) { > + /* ENOENT means the operation was interrupted, it's safe > + to restart */ > + if (err == ENOENT) > + goto restart; > + > + if (err == ENODEV) { > + /* Filesystem was unmounted, or connection was aborted > + via /sys/fs/fuse/connections */ > + fuse_session_exit(se); > + return 0; > + } > + /* Errors occurring during normal operation: EINTR (read > + interrupted), EAGAIN (nonblocking I/O), ENODEV (filesystem > + umounted) */ > + if (err != EINTR && err != EAGAIN) > + perror("fuse: reading device"); > + return -err; > + } > + if ((size_t) res < sizeof(struct fuse_in_header)) { > + fuse_log(FUSE_LOG_ERR, "short read on fuse device\n"); > + return -EIO; > + } > + > + buf->size = res; > + > + return res; > +} > + > +struct fuse_session *fuse_session_new(struct fuse_args *args, > + const struct fuse_lowlevel_ops *op, > + size_t op_size, void *userdata) > +{ > + int err; > + struct fuse_session *se; > + struct mount_opts *mo; > + > + if (sizeof(struct fuse_lowlevel_ops) < op_size) { > + fuse_log(FUSE_LOG_ERR, "fuse: warning: library too old, some > operations may not work\n"); > + op_size = sizeof(struct fuse_lowlevel_ops); > + } > + > + if (args->argc == 0) { > + fuse_log(FUSE_LOG_ERR, "fuse: empty argv passed to > fuse_session_new().\n"); > + return NULL; > + } > + > + se = (struct fuse_session *) calloc(1, sizeof(struct fuse_session)); > + if (se == NULL) { > + fuse_log(FUSE_LOG_ERR, "fuse: failed to allocate fuse > object\n"); > + goto out1; > + } > + se->fd = -1; > + se->conn.max_write = UINT_MAX; > + se->conn.max_readahead = UINT_MAX; > + > + /* Parse options */ > + if(fuse_opt_parse(args, se, fuse_ll_opts, NULL) == -1) > + goto out2; > + if(se->deny_others) { > + /* Allowing access only by root is done by instructing > + * kernel to allow access by everyone, and then restricting > + * access to root and mountpoint owner in libfuse. > + */ > + // We may be adding the option a second time, but > + // that doesn't hurt. > + if(fuse_opt_add_arg(args, "-oallow_other") == -1) > + goto out2; > + } > + mo = parse_mount_opts(args); > + if (mo == NULL) > + goto out3; > + > + if(args->argc == 1 && > + args->argv[0][0] == '-') { > + fuse_log(FUSE_LOG_ERR, "fuse: warning: argv[0] looks like an > option, but " > + "will be ignored\n"); > + } else if (args->argc != 1) { > + int i; > + fuse_log(FUSE_LOG_ERR, "fuse: unknown option(s): `"); > + for(i = 1; i < args->argc-1; i++) > + fuse_log(FUSE_LOG_ERR, "%s ", args->argv[i]); > + fuse_log(FUSE_LOG_ERR, "%s'\n", args->argv[i]); > + goto out4; > + } > + > + if (se->debug) > + fuse_log(FUSE_LOG_DEBUG, "FUSE library version: %s\n", > PACKAGE_VERSION); > + > + se->bufsize = FUSE_MAX_MAX_PAGES * getpagesize() + > + FUSE_BUFFER_HEADER_SIZE; > + > + list_init_req(&se->list); > + list_init_req(&se->interrupts); > + list_init_nreq(&se->notify_list); > + se->notify_ctr = 1; > + fuse_mutex_init(&se->lock); > + > + err = pthread_key_create(&se->pipe_key, fuse_ll_pipe_destructor); > + if (err) { > + fuse_log(FUSE_LOG_ERR, "fuse: failed to create thread specific > key: %s\n", > + strerror(err)); > + goto out5; > + } > + > + memcpy(&se->op, op, op_size); > + se->owner = getuid(); > + se->userdata = userdata; > + > + se->mo = mo; > + return se; > + > +out5: > + pthread_mutex_destroy(&se->lock); > +out4: > + fuse_opt_free_args(args); > +out3: > + free(mo); > +out2: > + free(se); > +out1: > + return NULL; > +} > + > +int fuse_session_mount(struct fuse_session *se, const char *mountpoint) > +{ > + int fd; > + > + /* > + * Make sure file descriptors 0, 1 and 2 are open, otherwise chaos > + * would ensue. > + */ > + do { > + fd = open("/dev/null", O_RDWR); > + if (fd > 2) > + close(fd); > + } while (fd >= 0 && fd <= 2); > + > + /* > + * To allow FUSE daemons to run without privileges, the caller may open > + * /dev/fuse before launching the file system and pass on the file > + * descriptor by specifying /dev/fd/N as the mount point. Note that the > + * parent process takes care of performing the mount in this case. > + */ > + fd = fuse_mnt_parse_fuse_fd(mountpoint); > + if (fd != -1) { > + if (fcntl(fd, F_GETFD) == -1) { > + fuse_log(FUSE_LOG_ERR, > + "fuse: Invalid file descriptor /dev/fd/%u\n", > + fd); > + return -1; > + } > + se->fd = fd; > + return 0; > + } > + > + /* Open channel */ > + fd = fuse_kern_mount(mountpoint, se->mo); > + if (fd == -1) > + return -1; > + se->fd = fd; > + > + /* Save mountpoint */ > + se->mountpoint = strdup(mountpoint); > + if (se->mountpoint == NULL) > + goto error_out; > + > + return 0; > + > +error_out: > + fuse_kern_unmount(mountpoint, fd); > + return -1; > +} > + > +int fuse_session_fd(struct fuse_session *se) > +{ > + return se->fd; > +} > + > +void fuse_session_unmount(struct fuse_session *se) > +{ > + if (se->mountpoint != NULL) { > + fuse_kern_unmount(se->mountpoint, se->fd); > + free(se->mountpoint); > + se->mountpoint = NULL; > + } > +} > + > +#ifdef linux > +int fuse_req_getgroups(fuse_req_t req, int size, gid_t list[]) > +{ > + char *buf; > + size_t bufsize = 1024; > + char path[128]; > + int ret; > + int fd; > + unsigned long pid = req->ctx.pid; > + char *s; > + > + sprintf(path, "/proc/%lu/task/%lu/status", pid, pid); > + > +retry: > + buf = malloc(bufsize); > + if (buf == NULL) > + return -ENOMEM; > + > + ret = -EIO; > + fd = open(path, O_RDONLY); > + if (fd == -1) > + goto out_free; > + > + ret = read(fd, buf, bufsize); > + close(fd); > + if (ret < 0) { > + ret = -EIO; > + goto out_free; > + } > + > + if ((size_t)ret == bufsize) { > + free(buf); > + bufsize *= 4; > + goto retry; > + } > + > + ret = -EIO; > + s = strstr(buf, "\nGroups:"); > + if (s == NULL) > + goto out_free; > + > + s += 8; > + ret = 0; > + while (1) { > + char *end; > + unsigned long val = strtoul(s, &end, 0); > + if (end == s) > + break; > + > + s = end; > + if (ret < size) > + list[ret] = val; > + ret++; > + } > + > +out_free: > + free(buf); > + return ret; > +} > +#else /* linux */ > +/* > + * This is currently not implemented on other than Linux... > + */ > +int fuse_req_getgroups(fuse_req_t req, int size, gid_t list[]) > +{ > + (void) req; (void) size; (void) list; > + return -ENOSYS; > +} > +#endif > + > +void fuse_session_exit(struct fuse_session *se) > +{ > + se->exited = 1; > +} > + > +void fuse_session_reset(struct fuse_session *se) > +{ > + se->exited = 0; > + se->error = 0; > +} > + > +int fuse_session_exited(struct fuse_session *se) > +{ > + return se->exited; > +} > -- > 2.23.0