提交 60e6679e 编写于 作者: T Theodore Ts'o

ext4: Drop whitespace at end of lines

This patch was generated using:

#!/usr/bin/perl -i
while (<>) {
    s/[ 	]+$//;
    print;
}
Signed-off-by: N"Theodore Ts'o" <tytso@mit.edu>
上级 4d92dc0f
...@@ -72,9 +72,9 @@ static int add_system_zone(struct ext4_sb_info *sbi, ...@@ -72,9 +72,9 @@ static int add_system_zone(struct ext4_sb_info *sbi,
else if (start_blk >= (entry->start_blk + entry->count)) else if (start_blk >= (entry->start_blk + entry->count))
n = &(*n)->rb_right; n = &(*n)->rb_right;
else { else {
if (start_blk + count > (entry->start_blk + if (start_blk + count > (entry->start_blk +
entry->count)) entry->count))
entry->count = (start_blk + count - entry->count = (start_blk + count -
entry->start_blk); entry->start_blk);
new_node = *n; new_node = *n;
new_entry = rb_entry(new_node, struct ext4_system_zone, new_entry = rb_entry(new_node, struct ext4_system_zone,
......
...@@ -75,7 +75,7 @@ typedef __u32 ext4_lblk_t; ...@@ -75,7 +75,7 @@ typedef __u32 ext4_lblk_t;
typedef unsigned int ext4_group_t; typedef unsigned int ext4_group_t;
/* /*
* Flags used in mballoc's allocation_context flags field. * Flags used in mballoc's allocation_context flags field.
* *
* Also used to show what's going on for debugging purposes when the * Also used to show what's going on for debugging purposes when the
* flag field is exported via the traceport interface * flag field is exported via the traceport interface
...@@ -470,7 +470,7 @@ struct ext4_new_group_data { ...@@ -470,7 +470,7 @@ struct ext4_new_group_data {
#define EXT4_GET_BLOCKS_CREATE_UNINIT_EXT (EXT4_GET_BLOCKS_UNINIT_EXT|\ #define EXT4_GET_BLOCKS_CREATE_UNINIT_EXT (EXT4_GET_BLOCKS_UNINIT_EXT|\
EXT4_GET_BLOCKS_CREATE) EXT4_GET_BLOCKS_CREATE)
/* Caller is from the delayed allocation writeout path, /* Caller is from the delayed allocation writeout path,
so set the magic i_delalloc_reserve_flag after taking the so set the magic i_delalloc_reserve_flag after taking the
inode allocation semaphore for */ inode allocation semaphore for */
#define EXT4_GET_BLOCKS_DELALLOC_RESERVE 0x0004 #define EXT4_GET_BLOCKS_DELALLOC_RESERVE 0x0004
/* caller is from the direct IO path, request to creation of an /* caller is from the direct IO path, request to creation of an
......
...@@ -182,10 +182,10 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode, ...@@ -182,10 +182,10 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode,
if (flex_size >= EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME) { if (flex_size >= EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME) {
/* /*
* If there are at least EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME * If there are at least EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME
* block groups per flexgroup, reserve the first block * block groups per flexgroup, reserve the first block
* group for directories and special files. Regular * group for directories and special files. Regular
* files will start at the second block group. This * files will start at the second block group. This
* tends to speed up directory access and improves * tends to speed up directory access and improves
* fsck times. * fsck times.
*/ */
block_group &= ~(flex_size-1); block_group &= ~(flex_size-1);
...@@ -2034,7 +2034,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block, ...@@ -2034,7 +2034,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block,
struct ext4_ext_cache *cex; struct ext4_ext_cache *cex;
int ret = EXT4_EXT_CACHE_NO; int ret = EXT4_EXT_CACHE_NO;
/* /*
* We borrow i_block_reservation_lock to protect i_cached_extent * We borrow i_block_reservation_lock to protect i_cached_extent
*/ */
spin_lock(&EXT4_I(inode)->i_block_reservation_lock); spin_lock(&EXT4_I(inode)->i_block_reservation_lock);
......
...@@ -66,7 +66,7 @@ int ext4_sync_file(struct file *file, struct dentry *dentry, int datasync) ...@@ -66,7 +66,7 @@ int ext4_sync_file(struct file *file, struct dentry *dentry, int datasync)
ret = flush_completed_IO(inode); ret = flush_completed_IO(inode);
if (ret < 0) if (ret < 0)
return ret; return ret;
if (!journal) if (!journal)
return simple_fsync(file, dentry, datasync); return simple_fsync(file, dentry, datasync);
......
...@@ -784,7 +784,7 @@ static int ext4_alloc_branch(handle_t *handle, struct inode *inode, ...@@ -784,7 +784,7 @@ static int ext4_alloc_branch(handle_t *handle, struct inode *inode,
/* Allocation failed, free what we already allocated */ /* Allocation failed, free what we already allocated */
ext4_free_blocks(handle, inode, 0, new_blocks[0], 1, 0); ext4_free_blocks(handle, inode, 0, new_blocks[0], 1, 0);
for (i = 1; i <= n ; i++) { for (i = 1; i <= n ; i++) {
/* /*
* branch[i].bh is newly allocated, so there is no * branch[i].bh is newly allocated, so there is no
* need to revoke the block, which is why we don't * need to revoke the block, which is why we don't
* need to set EXT4_FREE_BLOCKS_METADATA. * need to set EXT4_FREE_BLOCKS_METADATA.
...@@ -874,7 +874,7 @@ static int ext4_splice_branch(handle_t *handle, struct inode *inode, ...@@ -874,7 +874,7 @@ static int ext4_splice_branch(handle_t *handle, struct inode *inode,
err_out: err_out:
for (i = 1; i <= num; i++) { for (i = 1; i <= num; i++) {
/* /*
* branch[i].bh is newly allocated, so there is no * branch[i].bh is newly allocated, so there is no
* need to revoke the block, which is why we don't * need to revoke the block, which is why we don't
* need to set EXT4_FREE_BLOCKS_METADATA. * need to set EXT4_FREE_BLOCKS_METADATA.
......
...@@ -258,7 +258,7 @@ long ext4_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) ...@@ -258,7 +258,7 @@ long ext4_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
if (me.moved_len > 0) if (me.moved_len > 0)
file_remove_suid(donor_filp); file_remove_suid(donor_filp);
if (copy_to_user((struct move_extent __user *)arg, if (copy_to_user((struct move_extent __user *)arg,
&me, sizeof(me))) &me, sizeof(me)))
err = -EFAULT; err = -EFAULT;
mext_out: mext_out:
......
...@@ -2691,7 +2691,7 @@ int __init init_ext4_mballoc(void) ...@@ -2691,7 +2691,7 @@ int __init init_ext4_mballoc(void)
void exit_ext4_mballoc(void) void exit_ext4_mballoc(void)
{ {
/* /*
* Wait for completion of call_rcu()'s on ext4_pspace_cachep * Wait for completion of call_rcu()'s on ext4_pspace_cachep
* before destroying the slab cache. * before destroying the slab cache.
*/ */
...@@ -3330,7 +3330,7 @@ static void ext4_mb_put_pa(struct ext4_allocation_context *ac, ...@@ -3330,7 +3330,7 @@ static void ext4_mb_put_pa(struct ext4_allocation_context *ac,
spin_unlock(&pa->pa_lock); spin_unlock(&pa->pa_lock);
grp_blk = pa->pa_pstart; grp_blk = pa->pa_pstart;
/* /*
* If doing group-based preallocation, pa_pstart may be in the * If doing group-based preallocation, pa_pstart may be in the
* next group when pa is used up * next group when pa is used up
*/ */
...@@ -4534,12 +4534,12 @@ void ext4_free_blocks(handle_t *handle, struct inode *inode, ...@@ -4534,12 +4534,12 @@ void ext4_free_blocks(handle_t *handle, struct inode *inode,
if (!bh) if (!bh)
tbh = sb_find_get_block(inode->i_sb, tbh = sb_find_get_block(inode->i_sb,
block + i); block + i);
ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA, ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA,
inode, tbh, block + i); inode, tbh, block + i);
} }
} }
/* /*
* We need to make sure we don't reuse the freed block until * We need to make sure we don't reuse the freed block until
* after the transaction is committed, which we can do by * after the transaction is committed, which we can do by
* treating the block as metadata, below. We make an * treating the block as metadata, below. We make an
......
...@@ -187,7 +187,7 @@ unsigned int ext4_rec_len_from_disk(__le16 dlen, unsigned blocksize) ...@@ -187,7 +187,7 @@ unsigned int ext4_rec_len_from_disk(__le16 dlen, unsigned blocksize)
return blocksize; return blocksize;
return (len & 65532) | ((len & 3) << 16); return (len & 65532) | ((len & 3) << 16);
} }
__le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize) __le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
{ {
if ((len > blocksize) || (blocksize > (1 << 18)) || (len & 3)) if ((len > blocksize) || (blocksize > (1 << 18)) || (len & 3))
...@@ -197,7 +197,7 @@ __le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize) ...@@ -197,7 +197,7 @@ __le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
if (len == blocksize) { if (len == blocksize) {
if (blocksize == 65536) if (blocksize == 65536)
return cpu_to_le16(EXT4_MAX_REC_LEN); return cpu_to_le16(EXT4_MAX_REC_LEN);
else else
return cpu_to_le16(0); return cpu_to_le16(0);
} }
return cpu_to_le16((len & 65532) | ((len >> 16) & 3)); return cpu_to_le16((len & 65532) | ((len >> 16) & 3));
...@@ -349,7 +349,7 @@ struct stats dx_show_entries(struct dx_hash_info *hinfo, struct inode *dir, ...@@ -349,7 +349,7 @@ struct stats dx_show_entries(struct dx_hash_info *hinfo, struct inode *dir,
brelse(bh); brelse(bh);
} }
if (bcount) if (bcount)
printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n", printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n",
levels ? "" : " ", names, space/bcount, levels ? "" : " ", names, space/bcount,
(space/bcount)*100/blocksize); (space/bcount)*100/blocksize);
return (struct stats) { names, space, bcount}; return (struct stats) { names, space, bcount};
...@@ -653,7 +653,7 @@ int ext4_htree_fill_tree(struct file *dir_file, __u32 start_hash, ...@@ -653,7 +653,7 @@ int ext4_htree_fill_tree(struct file *dir_file, __u32 start_hash,
int ret, err; int ret, err;
__u32 hashval; __u32 hashval;
dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n", dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n",
start_hash, start_minor_hash)); start_hash, start_minor_hash));
dir = dir_file->f_path.dentry->d_inode; dir = dir_file->f_path.dentry->d_inode;
if (!(ext4_test_inode_flag(dir, EXT4_INODE_INDEX))) { if (!(ext4_test_inode_flag(dir, EXT4_INODE_INDEX))) {
...@@ -1141,7 +1141,7 @@ dx_move_dirents(char *from, char *to, struct dx_map_entry *map, int count, ...@@ -1141,7 +1141,7 @@ dx_move_dirents(char *from, char *to, struct dx_map_entry *map, int count,
unsigned rec_len = 0; unsigned rec_len = 0;
while (count--) { while (count--) {
struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *) struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *)
(from + (map->offs<<2)); (from + (map->offs<<2));
rec_len = EXT4_DIR_REC_LEN(de->name_len); rec_len = EXT4_DIR_REC_LEN(de->name_len);
memcpy (to, de, rec_len); memcpy (to, de, rec_len);
......
...@@ -2216,7 +2216,7 @@ static unsigned long ext4_get_stripe_size(struct ext4_sb_info *sbi) ...@@ -2216,7 +2216,7 @@ static unsigned long ext4_get_stripe_size(struct ext4_sb_info *sbi)
struct ext4_attr { struct ext4_attr {
struct attribute attr; struct attribute attr;
ssize_t (*show)(struct ext4_attr *, struct ext4_sb_info *, char *); ssize_t (*show)(struct ext4_attr *, struct ext4_sb_info *, char *);
ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *, ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *,
const char *, size_t); const char *, size_t);
int offset; int offset;
}; };
...@@ -3383,7 +3383,7 @@ static int ext4_commit_super(struct super_block *sb, int sync) ...@@ -3383,7 +3383,7 @@ static int ext4_commit_super(struct super_block *sb, int sync)
if (!(sb->s_flags & MS_RDONLY)) if (!(sb->s_flags & MS_RDONLY))
es->s_wtime = cpu_to_le32(get_seconds()); es->s_wtime = cpu_to_le32(get_seconds());
es->s_kbytes_written = es->s_kbytes_written =
cpu_to_le64(EXT4_SB(sb)->s_kbytes_written + cpu_to_le64(EXT4_SB(sb)->s_kbytes_written +
((part_stat_read(sb->s_bdev->bd_part, sectors[1]) - ((part_stat_read(sb->s_bdev->bd_part, sectors[1]) -
EXT4_SB(sb)->s_sectors_written_start) >> 1)); EXT4_SB(sb)->s_sectors_written_start) >> 1));
ext4_free_blocks_count_set(es, percpu_counter_sum_positive( ext4_free_blocks_count_set(es, percpu_counter_sum_positive(
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册