/* * (C) 2001 Clemson University and The University of Chicago * * Changes by Acxiom Corporation to add protocol version to kernel * communication, Copyright Acxiom Corporation, 2005. * * See COPYING in top-level directory. */ #include "protocol.h" #include "pvfs2-kernel.h" #include "pvfs2-dev-proto.h" #include "pvfs2-bufmap.h" #include #include /* this file implements the /dev/pvfs2-req device node */ static int open_access_count; #define DUMP_DEVICE_ERROR() \ do { \ gossip_err("*****************************************************\n");\ gossip_err("PVFS2 Device Error: You cannot open the device file "); \ gossip_err("\n/dev/%s more than once. Please make sure that\nthere " \ "are no ", PVFS2_REQDEVICE_NAME); \ gossip_err("instances of a program using this device\ncurrently " \ "running. (You must verify this!)\n"); \ gossip_err("For example, you can use the lsof program as follows:\n");\ gossip_err("'lsof | grep %s' (run this as root)\n", \ PVFS2_REQDEVICE_NAME); \ gossip_err(" open_access_count = %d\n", open_access_count); \ gossip_err("*****************************************************\n");\ } while (0) static int hash_func(__u64 tag, int table_size) { return do_div(tag, (unsigned int)table_size); } static void pvfs2_devreq_add_op(struct pvfs2_kernel_op_s *op) { int index = hash_func(op->tag, hash_table_size); spin_lock(&htable_ops_in_progress_lock); list_add_tail(&op->list, &htable_ops_in_progress[index]); spin_unlock(&htable_ops_in_progress_lock); } static struct pvfs2_kernel_op_s *pvfs2_devreq_remove_op(__u64 tag) { struct pvfs2_kernel_op_s *op, *next; int index; index = hash_func(tag, hash_table_size); spin_lock(&htable_ops_in_progress_lock); list_for_each_entry_safe(op, next, &htable_ops_in_progress[index], list) { if (op->tag == tag) { list_del(&op->list); spin_unlock(&htable_ops_in_progress_lock); return op; } } spin_unlock(&htable_ops_in_progress_lock); return NULL; } static int pvfs2_devreq_open(struct inode *inode, struct file *file) { int ret = -EINVAL; if (!(file->f_flags & O_NONBLOCK)) { gossip_err("pvfs2: device cannot be opened in blocking mode\n"); goto out; } ret = -EACCES; gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2-client-core: opening device\n"); mutex_lock(&devreq_mutex); if (open_access_count == 0) { ret = generic_file_open(inode, file); if (ret == 0) open_access_count++; } else { DUMP_DEVICE_ERROR(); } mutex_unlock(&devreq_mutex); out: gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2-client-core: open device complete (ret = %d)\n", ret); return ret; } static ssize_t pvfs2_devreq_read(struct file *file, char __user *buf, size_t count, loff_t *offset) { int ret = 0; ssize_t len = 0; struct pvfs2_kernel_op_s *cur_op = NULL; static __s32 magic = PVFS2_DEVREQ_MAGIC; __s32 proto_ver = PVFS_KERNEL_PROTO_VERSION; if (!(file->f_flags & O_NONBLOCK)) { /* We do not support blocking reads/opens any more */ gossip_err("pvfs2: blocking reads are not supported! (pvfs2-client-core bug)\n"); return -EINVAL; } else { struct pvfs2_kernel_op_s *op = NULL, *temp = NULL; /* get next op (if any) from top of list */ spin_lock(&pvfs2_request_list_lock); list_for_each_entry_safe(op, temp, &pvfs2_request_list, list) { __s32 fsid = fsid_of_op(op); /* * Check if this op's fsid is known and needs * remounting */ if (fsid != PVFS_FS_ID_NULL && fs_mount_pending(fsid) == 1) { gossip_debug(GOSSIP_DEV_DEBUG, "Skipping op tag %llu %s\n", llu(op->tag), get_opname_string(op)); continue; } else { /* * op does not belong to any particular fsid * or already mounted.. let it through */ cur_op = op; spin_lock(&cur_op->lock); list_del(&cur_op->list); cur_op->op_linger_tmp--; /* * if there is a trailer, re-add it to * the request list. */ if (cur_op->op_linger == 2 && cur_op->op_linger_tmp == 1) { if (cur_op->upcall.trailer_size <= 0 || cur_op->upcall.trailer_buf == NULL) gossip_err("BUG:trailer_size is %ld and trailer buf is %p\n", (long)cur_op->upcall.trailer_size, cur_op->upcall.trailer_buf); /* re-add it to the head of the list */ list_add(&cur_op->list, &pvfs2_request_list); } spin_unlock(&cur_op->lock); break; } } spin_unlock(&pvfs2_request_list_lock); } if (cur_op) { spin_lock(&cur_op->lock); gossip_debug(GOSSIP_DEV_DEBUG, "client-core: reading op tag %llu %s\n", llu(cur_op->tag), get_opname_string(cur_op)); if (op_state_in_progress(cur_op) || op_state_serviced(cur_op)) { if (cur_op->op_linger == 1) gossip_err("WARNING: Current op already queued...skipping\n"); } else if (cur_op->op_linger == 1 || (cur_op->op_linger == 2 && cur_op->op_linger_tmp == 0)) { /* * atomically move the operation to the * htable_ops_in_progress */ set_op_state_inprogress(cur_op); pvfs2_devreq_add_op(cur_op); } spin_unlock(&cur_op->lock); /* 2 cases * a) OPs with no trailers * b) OPs with trailers, Stage 1 * Either way push the upcall out */ if (cur_op->op_linger == 1 || (cur_op->op_linger == 2 && cur_op->op_linger_tmp == 1)) { len = MAX_ALIGNED_DEV_REQ_UPSIZE; if ((size_t) len <= count) { ret = copy_to_user(buf, &proto_ver, sizeof(__s32)); if (ret == 0) { ret = copy_to_user(buf + sizeof(__s32), &magic, sizeof(__s32)); if (ret == 0) { ret = copy_to_user(buf+2 * sizeof(__s32), &cur_op->tag, sizeof(__u64)); if (ret == 0) { ret = copy_to_user( buf + 2 * sizeof(__s32) + sizeof(__u64), &cur_op->upcall, sizeof(struct pvfs2_upcall_s)); } } } if (ret) { gossip_err("Failed to copy data to user space\n"); len = -EFAULT; } } else { gossip_err ("Failed to copy data to user space\n"); len = -EIO; } } /* Stage 2: Push the trailer out */ else if (cur_op->op_linger == 2 && cur_op->op_linger_tmp == 0) { len = cur_op->upcall.trailer_size; if ((size_t) len <= count) { ret = copy_to_user(buf, cur_op->upcall.trailer_buf, len); if (ret) { gossip_err("Failed to copy trailer to user space\n"); len = -EFAULT; } } else { gossip_err("Read buffer for trailer is too small (%ld as opposed to %ld)\n", (long)count, (long)len); len = -EIO; } } else { gossip_err("cur_op: %p (op_linger %d), (op_linger_tmp %d), erroneous request list?\n", cur_op, cur_op->op_linger, cur_op->op_linger_tmp); len = 0; } } else if (file->f_flags & O_NONBLOCK) { /* * if in non-blocking mode, return EAGAIN since no requests are * ready yet */ len = -EAGAIN; } return len; } /* Function for writev() callers into the device */ static ssize_t pvfs2_devreq_writev(struct file *file, const struct iovec *iov, size_t count, loff_t *offset) { struct pvfs2_kernel_op_s *op = NULL; void *buffer = NULL; void *ptr = NULL; unsigned long i = 0; static int max_downsize = MAX_ALIGNED_DEV_REQ_DOWNSIZE; int ret = 0, num_remaining = max_downsize; int notrailer_count = 4; /* num elements in iovec without trailer */ int payload_size = 0; __s32 magic = 0; __s32 proto_ver = 0; __u64 tag = 0; ssize_t total_returned_size = 0; /* Either there is a trailer or there isn't */ if (count != notrailer_count && count != (notrailer_count + 1)) { gossip_err("Error: Number of iov vectors is (%zu) and notrailer count is %d\n", count, notrailer_count); return -EPROTO; } buffer = dev_req_alloc(); if (!buffer) return -ENOMEM; ptr = buffer; for (i = 0; i < notrailer_count; i++) { if (iov[i].iov_len > num_remaining) { gossip_err ("writev error: Freeing buffer and returning\n"); dev_req_release(buffer); return -EMSGSIZE; } ret = copy_from_user(ptr, iov[i].iov_base, iov[i].iov_len); if (ret) { gossip_err("Failed to copy data from user space\n"); dev_req_release(buffer); return -EIO; } num_remaining -= iov[i].iov_len; ptr += iov[i].iov_len; payload_size += iov[i].iov_len; } total_returned_size = payload_size; /* these elements are currently 8 byte aligned (8 bytes for (version + * magic) 8 bytes for tag). If you add another element, either * make it 8 bytes big, or use get_unaligned when asigning. */ ptr = buffer; proto_ver = *((__s32 *) ptr); ptr += sizeof(__s32); magic = *((__s32 *) ptr); ptr += sizeof(__s32); tag = *((__u64 *) ptr); ptr += sizeof(__u64); if (magic != PVFS2_DEVREQ_MAGIC) { gossip_err("Error: Device magic number does not match.\n"); dev_req_release(buffer); return -EPROTO; } /* * proto_ver = 20902 for 2.9.2 */ op = pvfs2_devreq_remove_op(tag); if (op) { /* Increase ref count! */ get_op(op); /* cut off magic and tag from payload size */ payload_size -= (2 * sizeof(__s32) + sizeof(__u64)); if (payload_size <= sizeof(struct pvfs2_downcall_s)) /* copy the passed in downcall into the op */ memcpy(&op->downcall, ptr, sizeof(struct pvfs2_downcall_s)); else gossip_debug(GOSSIP_DEV_DEBUG, "writev: Ignoring %d bytes\n", payload_size); /* Do not allocate needlessly if client-core forgets * to reset trailer size on op errors. */ if (op->downcall.status == 0 && op->downcall.trailer_size > 0) { gossip_debug(GOSSIP_DEV_DEBUG, "writev: trailer size %ld\n", (unsigned long)op->downcall.trailer_size); if (count != (notrailer_count + 1)) { gossip_err("Error: trailer size (%ld) is non-zero, no trailer elements though? (%zu)\n", (unsigned long)op->downcall.trailer_size, count); dev_req_release(buffer); put_op(op); return -EPROTO; } if (iov[notrailer_count].iov_len > op->downcall.trailer_size) { gossip_err("writev error: trailer size (%ld) != iov_len (%ld)\n", (unsigned long)op->downcall.trailer_size, (unsigned long)iov[notrailer_count].iov_len); dev_req_release(buffer); put_op(op); return -EMSGSIZE; } /* Allocate a buffer large enough to hold the * trailer bytes. */ op->downcall.trailer_buf = vmalloc(op->downcall.trailer_size); if (op->downcall.trailer_buf != NULL) { gossip_debug(GOSSIP_DEV_DEBUG, "vmalloc: %p\n", op->downcall.trailer_buf); ret = copy_from_user(op->downcall.trailer_buf, iov[notrailer_count]. iov_base, iov[notrailer_count]. iov_len); if (ret) { gossip_err("Failed to copy trailer data from user space\n"); dev_req_release(buffer); gossip_debug(GOSSIP_DEV_DEBUG, "vfree: %p\n", op->downcall.trailer_buf); vfree(op->downcall.trailer_buf); op->downcall.trailer_buf = NULL; put_op(op); return -EIO; } } else { /* Change downcall status */ op->downcall.status = -ENOMEM; gossip_err("writev: could not vmalloc for trailer!\n"); } } /* if this operation is an I/O operation and if it was * initiated on behalf of a *synchronous* VFS I/O operation, * only then we need to wait * for all data to be copied before we can return to avoid * buffer corruption and races that can pull the buffers * out from under us. * * Essentially we're synchronizing with other parts of the * vfs implicitly by not allowing the user space * application reading/writing this device to return until * the buffers are done being used. */ if ((op->upcall.type == PVFS2_VFS_OP_FILE_IO && op->upcall.req.io.async_vfs_io == PVFS_VFS_SYNC_IO) || op->upcall.type == PVFS2_VFS_OP_FILE_IOX) { int timed_out = 0; DECLARE_WAITQUEUE(wait_entry, current); /* tell the vfs op waiting on a waitqueue * that this op is done */ spin_lock(&op->lock); set_op_state_serviced(op); spin_unlock(&op->lock); add_wait_queue_exclusive(&op->io_completion_waitq, &wait_entry); wake_up_interruptible(&op->waitq); while (1) { set_current_state(TASK_INTERRUPTIBLE); spin_lock(&op->lock); if (op->io_completed) { spin_unlock(&op->lock); break; } spin_unlock(&op->lock); if (!signal_pending(current)) { int timeout = MSECS_TO_JIFFIES(1000 * op_timeout_secs); if (!schedule_timeout(timeout)) { gossip_debug(GOSSIP_DEV_DEBUG, "*** I/O wait time is up\n"); timed_out = 1; break; } continue; } gossip_debug(GOSSIP_DEV_DEBUG, "*** signal on I/O wait -- aborting\n"); break; } set_current_state(TASK_RUNNING); remove_wait_queue(&op->io_completion_waitq, &wait_entry); /* NOTE: for I/O operations we handle releasing the op * object except in the case of timeout. the reason we * can't free the op in timeout cases is that the op * service logic in the vfs retries operations using * the same op ptr, thus it can't be freed. */ if (!timed_out) op_release(op); } else { /* * tell the vfs op waiting on a waitqueue that * this op is done */ spin_lock(&op->lock); set_op_state_serviced(op); spin_unlock(&op->lock); /* for every other operation (i.e. non-I/O), we need to wake up the callers for downcall completion notification */ wake_up_interruptible(&op->waitq); } } else { /* ignore downcalls that we're not interested in */ gossip_debug(GOSSIP_DEV_DEBUG, "WARNING: No one's waiting for tag %llu\n", llu(tag)); } dev_req_release(buffer); return total_returned_size; } static ssize_t pvfs2_devreq_write_iter(struct kiocb *iocb, struct iov_iter *iter) { return pvfs2_devreq_writev(iocb->ki_filp, iter->iov, iter->nr_segs, &iocb->ki_pos); } /* Returns whether any FS are still pending remounted */ static int mark_all_pending_mounts(void) { int unmounted = 1; struct pvfs2_sb_info_s *pvfs2_sb = NULL; spin_lock(&pvfs2_superblocks_lock); list_for_each_entry(pvfs2_sb, &pvfs2_superblocks, list) { /* All of these file system require a remount */ pvfs2_sb->mount_pending = 1; unmounted = 0; } spin_unlock(&pvfs2_superblocks_lock); return unmounted; } /* * Determine if a given file system needs to be remounted or not * Returns -1 on error * 0 if already mounted * 1 if needs remount */ int fs_mount_pending(__s32 fsid) { int mount_pending = -1; struct pvfs2_sb_info_s *pvfs2_sb = NULL; spin_lock(&pvfs2_superblocks_lock); list_for_each_entry(pvfs2_sb, &pvfs2_superblocks, list) { if (pvfs2_sb->fs_id == fsid) { mount_pending = pvfs2_sb->mount_pending; break; } } spin_unlock(&pvfs2_superblocks_lock); return mount_pending; } /* * NOTE: gets called when the last reference to this device is dropped. * Using the open_access_count variable, we enforce a reference count * on this file so that it can be opened by only one process at a time. * the devreq_mutex is used to make sure all i/o has completed * before we call pvfs_bufmap_finalize, and similar such tricky * situations */ static int pvfs2_devreq_release(struct inode *inode, struct file *file) { int unmounted = 0; gossip_debug(GOSSIP_DEV_DEBUG, "%s:pvfs2-client-core: exiting, closing device\n", __func__); mutex_lock(&devreq_mutex); pvfs_bufmap_finalize(); open_access_count--; unmounted = mark_all_pending_mounts(); gossip_debug(GOSSIP_DEV_DEBUG, "PVFS2 Device Close: Filesystem(s) %s\n", (unmounted ? "UNMOUNTED" : "MOUNTED")); mutex_unlock(&devreq_mutex); /* * Walk through the list of ops in the request list, mark them * as purged and wake them up. */ purge_waiting_ops(); /* * Walk through the hash table of in progress operations; mark * them as purged and wake them up */ purge_inprogress_ops(); gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2-client-core: device close complete\n"); return 0; } int is_daemon_in_service(void) { int in_service; /* * What this function does is checks if client-core is alive * based on the access count we maintain on the device. */ mutex_lock(&devreq_mutex); in_service = open_access_count == 1 ? 0 : -EIO; mutex_unlock(&devreq_mutex); return in_service; } static inline long check_ioctl_command(unsigned int command) { /* Check for valid ioctl codes */ if (_IOC_TYPE(command) != PVFS_DEV_MAGIC) { gossip_err("device ioctl magic numbers don't match! Did you rebuild pvfs2-client-core/libpvfs2? [cmd %x, magic %x != %x]\n", command, _IOC_TYPE(command), PVFS_DEV_MAGIC); return -EINVAL; } /* and valid ioctl commands */ if (_IOC_NR(command) >= PVFS_DEV_MAXNR || _IOC_NR(command) <= 0) { gossip_err("Invalid ioctl command number [%d >= %d]\n", _IOC_NR(command), PVFS_DEV_MAXNR); return -ENOIOCTLCMD; } return 0; } static long dispatch_ioctl_command(unsigned int command, unsigned long arg) { static __s32 magic = PVFS2_DEVREQ_MAGIC; static __s32 max_up_size = MAX_ALIGNED_DEV_REQ_UPSIZE; static __s32 max_down_size = MAX_ALIGNED_DEV_REQ_DOWNSIZE; struct PVFS_dev_map_desc user_desc; int ret = 0; struct dev_mask_info_s mask_info = { 0 }; struct dev_mask2_info_s mask2_info = { 0, 0 }; int upstream_kmod = 1; struct list_head *tmp = NULL; struct pvfs2_sb_info_s *pvfs2_sb = NULL; /* mtmoore: add locking here */ switch (command) { case PVFS_DEV_GET_MAGIC: return ((put_user(magic, (__s32 __user *) arg) == -EFAULT) ? -EIO : 0); case PVFS_DEV_GET_MAX_UPSIZE: return ((put_user(max_up_size, (__s32 __user *) arg) == -EFAULT) ? -EIO : 0); case PVFS_DEV_GET_MAX_DOWNSIZE: return ((put_user(max_down_size, (__s32 __user *) arg) == -EFAULT) ? -EIO : 0); case PVFS_DEV_MAP: ret = copy_from_user(&user_desc, (struct PVFS_dev_map_desc __user *) arg, sizeof(struct PVFS_dev_map_desc)); return ret ? -EIO : pvfs_bufmap_initialize(&user_desc); case PVFS_DEV_REMOUNT_ALL: gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2_devreq_ioctl: got PVFS_DEV_REMOUNT_ALL\n"); /* * remount all mounted pvfs2 volumes to regain the lost * dynamic mount tables (if any) -- NOTE: this is done * without keeping the superblock list locked due to the * upcall/downcall waiting. also, the request semaphore is * used to ensure that no operations will be serviced until * all of the remounts are serviced (to avoid ops between * mounts to fail) */ ret = mutex_lock_interruptible(&request_mutex); if (ret < 0) return ret; gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2_devreq_ioctl: priority remount in progress\n"); list_for_each(tmp, &pvfs2_superblocks) { pvfs2_sb = list_entry(tmp, struct pvfs2_sb_info_s, list); if (pvfs2_sb && (pvfs2_sb->sb)) { gossip_debug(GOSSIP_DEV_DEBUG, "Remounting SB %p\n", pvfs2_sb); ret = pvfs2_remount(pvfs2_sb->sb); if (ret) { gossip_debug(GOSSIP_DEV_DEBUG, "SB %p remount failed\n", pvfs2_sb); break; } } } gossip_debug(GOSSIP_DEV_DEBUG, "pvfs2_devreq_ioctl: priority remount complete\n"); mutex_unlock(&request_mutex); return ret; case PVFS_DEV_UPSTREAM: ret = copy_to_user((void __user *)arg, &upstream_kmod, sizeof(upstream_kmod)); if (ret != 0) return -EIO; else return ret; case PVFS_DEV_CLIENT_MASK: ret = copy_from_user(&mask2_info, (void __user *)arg, sizeof(struct dev_mask2_info_s)); if (ret != 0) return -EIO; client_debug_mask.mask1 = mask2_info.mask1_value; client_debug_mask.mask2 = mask2_info.mask2_value; pr_info("%s: client debug mask has been been received " ":%llx: :%llx:\n", __func__, (unsigned long long)client_debug_mask.mask1, (unsigned long long)client_debug_mask.mask2); return ret; case PVFS_DEV_CLIENT_STRING: ret = copy_from_user(&client_debug_array_string, (void __user *)arg, PVFS2_MAX_DEBUG_STRING_LEN); if (ret != 0) { pr_info("%s: " "PVFS_DEV_CLIENT_STRING: copy_from_user failed" "\n", __func__); return -EIO; } pr_info("%s: client debug array string has been been received." "\n", __func__); if (!help_string_initialized) { /* Free the "we don't know yet" default string... */ kfree(debug_help_string); /* build a proper debug help string */ if (orangefs_prepare_debugfs_help_string(0)) { gossip_err("%s: " "prepare_debugfs_help_string failed" "\n", __func__); return -EIO; } /* Replace the boilerplate boot-time debug-help file. */ debugfs_remove(help_file_dentry); help_file_dentry = debugfs_create_file( ORANGEFS_KMOD_DEBUG_HELP_FILE, 0444, debug_dir, debug_help_string, &debug_help_fops); if (!help_file_dentry) { gossip_err("%s: debugfs_create_file failed for" " :%s:!\n", __func__, ORANGEFS_KMOD_DEBUG_HELP_FILE); return -EIO; } } debug_mask_to_string(&client_debug_mask, 1); debugfs_remove(client_debug_dentry); pvfs2_client_debug_init(); help_string_initialized++; return ret; case PVFS_DEV_DEBUG: ret = copy_from_user(&mask_info, (void __user *)arg, sizeof(mask_info)); if (ret != 0) return -EIO; if (mask_info.mask_type == KERNEL_MASK) { if ((mask_info.mask_value == 0) && (kernel_mask_set_mod_init)) { /* * the kernel debug mask was set when the * kernel module was loaded; don't override * it if the client-core was started without * a value for PVFS2_KMODMASK. */ return 0; } debug_mask_to_string(&mask_info.mask_value, mask_info.mask_type); gossip_debug_mask = mask_info.mask_value; pr_info("PVFS: kernel debug mask has been modified to " ":%s: :%llx:\n", kernel_debug_string, (unsigned long long)gossip_debug_mask); } else if (mask_info.mask_type == CLIENT_MASK) { debug_mask_to_string(&mask_info.mask_value, mask_info.mask_type); pr_info("PVFS: client debug mask has been modified to" ":%s: :%llx:\n", client_debug_string, llu(mask_info.mask_value)); } else { gossip_lerr("Invalid mask type....\n"); return -EINVAL; } return ret; default: return -ENOIOCTLCMD; } return -ENOIOCTLCMD; } static long pvfs2_devreq_ioctl(struct file *file, unsigned int command, unsigned long arg) { long ret; /* Check for properly constructed commands */ ret = check_ioctl_command(command); if (ret < 0) return (int)ret; return (int)dispatch_ioctl_command(command, arg); } #ifdef CONFIG_COMPAT /* CONFIG_COMPAT is in .config */ /* Compat structure for the PVFS_DEV_MAP ioctl */ struct PVFS_dev_map_desc32 { compat_uptr_t ptr; __s32 total_size; __s32 size; __s32 count; }; static unsigned long translate_dev_map26(unsigned long args, long *error) { struct PVFS_dev_map_desc32 __user *p32 = (void __user *)args; /* * Depending on the architecture, allocate some space on the * user-call-stack based on our expected layout. */ struct PVFS_dev_map_desc __user *p = compat_alloc_user_space(sizeof(*p)); u32 addr; *error = 0; /* get the ptr from the 32 bit user-space */ if (get_user(addr, &p32->ptr)) goto err; /* try to put that into a 64-bit layout */ if (put_user(compat_ptr(addr), &p->ptr)) goto err; /* copy the remaining fields */ if (copy_in_user(&p->total_size, &p32->total_size, sizeof(__s32))) goto err; if (copy_in_user(&p->size, &p32->size, sizeof(__s32))) goto err; if (copy_in_user(&p->count, &p32->count, sizeof(__s32))) goto err; return (unsigned long)p; err: *error = -EFAULT; return 0; } /* * 32 bit user-space apps' ioctl handlers when kernel modules * is compiled as a 64 bit one */ static long pvfs2_devreq_compat_ioctl(struct file *filp, unsigned int cmd, unsigned long args) { long ret; unsigned long arg = args; /* Check for properly constructed commands */ ret = check_ioctl_command(cmd); if (ret < 0) return ret; if (cmd == PVFS_DEV_MAP) { /* * convert the arguments to what we expect internally * in kernel space */ arg = translate_dev_map26(args, &ret); if (ret < 0) { gossip_err("Could not translate dev map\n"); return ret; } } /* no other ioctl requires translation */ return dispatch_ioctl_command(cmd, arg); } #endif /* CONFIG_COMPAT is in .config */ /* * The following two ioctl32 functions had been refactored into the above * CONFIG_COMPAT ifdef, but that was an over simplification that was * not noticed until we tried to compile on power pc... */ #if (defined(CONFIG_COMPAT) && !defined(HAVE_REGISTER_IOCTL32_CONVERSION)) || !defined(CONFIG_COMPAT) static int pvfs2_ioctl32_init(void) { return 0; } static void pvfs2_ioctl32_cleanup(void) { return; } #endif /* the assigned character device major number */ static int pvfs2_dev_major; /* * Initialize pvfs2 device specific state: * Must be called at module load time only */ int pvfs2_dev_init(void) { int ret; /* register the ioctl32 sub-system */ ret = pvfs2_ioctl32_init(); if (ret < 0) return ret; /* register pvfs2-req device */ pvfs2_dev_major = register_chrdev(0, PVFS2_REQDEVICE_NAME, &pvfs2_devreq_file_operations); if (pvfs2_dev_major < 0) { gossip_debug(GOSSIP_DEV_DEBUG, "Failed to register /dev/%s (error %d)\n", PVFS2_REQDEVICE_NAME, pvfs2_dev_major); pvfs2_ioctl32_cleanup(); return pvfs2_dev_major; } gossip_debug(GOSSIP_DEV_DEBUG, "*** /dev/%s character device registered ***\n", PVFS2_REQDEVICE_NAME); gossip_debug(GOSSIP_DEV_DEBUG, "'mknod /dev/%s c %d 0'.\n", PVFS2_REQDEVICE_NAME, pvfs2_dev_major); return 0; } void pvfs2_dev_cleanup(void) { unregister_chrdev(pvfs2_dev_major, PVFS2_REQDEVICE_NAME); gossip_debug(GOSSIP_DEV_DEBUG, "*** /dev/%s character device unregistered ***\n", PVFS2_REQDEVICE_NAME); /* unregister the ioctl32 sub-system */ pvfs2_ioctl32_cleanup(); } static unsigned int pvfs2_devreq_poll(struct file *file, struct poll_table_struct *poll_table) { int poll_revent_mask = 0; if (open_access_count == 1) { poll_wait(file, &pvfs2_request_list_waitq, poll_table); spin_lock(&pvfs2_request_list_lock); if (!list_empty(&pvfs2_request_list)) poll_revent_mask |= POLL_IN; spin_unlock(&pvfs2_request_list_lock); } return poll_revent_mask; } const struct file_operations pvfs2_devreq_file_operations = { .owner = THIS_MODULE, .read = pvfs2_devreq_read, .write_iter = pvfs2_devreq_write_iter, .open = pvfs2_devreq_open, .release = pvfs2_devreq_release, .unlocked_ioctl = pvfs2_devreq_ioctl, #ifdef CONFIG_COMPAT /* CONFIG_COMPAT is in .config */ .compat_ioctl = pvfs2_devreq_compat_ioctl, #endif .poll = pvfs2_devreq_poll };