Loading fs/btrfs/check-integrity.c +1 −1 Original line number Diff line number Diff line Loading @@ -178,7 +178,7 @@ struct btrfsic_block { * Elements of this type are allocated dynamically and required because * each block object can refer to and can be ref from multiple blocks. * The key to lookup them in the hashtable is the dev_bytenr of * the block ref to plus the one from the block refered from. * the block ref to plus the one from the block referred from. * The fact that they are searchable via a hashtable and that a * ref_cnt is maintained is not required for the btrfs integrity * check algorithm itself, it is only used to make the output more Loading fs/btrfs/ctree.h +5 −5 Original line number Diff line number Diff line Loading @@ -1219,10 +1219,10 @@ struct btrfs_space_info { * we've called update_block_group and dropped the bytes_used counter * and increased the bytes_pinned counter. However this means that * bytes_pinned does not reflect the bytes that will be pinned once the * delayed refs are flushed, so this counter is inc'ed everytime we call * btrfs_free_extent so it is a realtime count of what will be freed * once the transaction is committed. It will be zero'ed everytime the * transaction commits. * delayed refs are flushed, so this counter is inc'ed every time we * call btrfs_free_extent so it is a realtime count of what will be * freed once the transaction is committed. It will be zero'ed every * time the transaction commits. */ struct percpu_counter total_bytes_pinned; Loading fs/btrfs/dev-replace.c +1 −1 Original line number Diff line number Diff line Loading @@ -858,7 +858,7 @@ int btrfs_dev_replace_is_ongoing(struct btrfs_dev_replace *dev_replace) * not called and the the filesystem is remounted * in degraded state. This does not stop the * dev_replace procedure. It needs to be canceled * manually if the cancelation is wanted. * manually if the cancellation is wanted. */ break; } Loading fs/btrfs/disk-io.c +1 −1 Original line number Diff line number Diff line Loading @@ -816,7 +816,7 @@ static void run_one_async_done(struct btrfs_work *work) waitqueue_active(&fs_info->async_submit_wait)) wake_up(&fs_info->async_submit_wait); /* If an error occured we just want to clean up the bio and move on */ /* If an error occurred we just want to clean up the bio and move on */ if (async->error) { async->bio->bi_error = async->error; bio_endio(async->bio); Loading fs/btrfs/extent-tree.c +1 −1 Original line number Diff line number Diff line Loading @@ -5758,7 +5758,7 @@ int btrfs_delalloc_reserve_metadata(struct inode *inode, u64 num_bytes) /* * This is tricky, but first we need to figure out how much we * free'd from any free-ers that occured during this * free'd from any free-ers that occurred during this * reservation, so we reset ->csum_bytes to the csum_bytes * before we dropped our lock, and then call the free for the * number of bytes that were freed while we were trying our Loading Loading
fs/btrfs/check-integrity.c +1 −1 Original line number Diff line number Diff line Loading @@ -178,7 +178,7 @@ struct btrfsic_block { * Elements of this type are allocated dynamically and required because * each block object can refer to and can be ref from multiple blocks. * The key to lookup them in the hashtable is the dev_bytenr of * the block ref to plus the one from the block refered from. * the block ref to plus the one from the block referred from. * The fact that they are searchable via a hashtable and that a * ref_cnt is maintained is not required for the btrfs integrity * check algorithm itself, it is only used to make the output more Loading
fs/btrfs/ctree.h +5 −5 Original line number Diff line number Diff line Loading @@ -1219,10 +1219,10 @@ struct btrfs_space_info { * we've called update_block_group and dropped the bytes_used counter * and increased the bytes_pinned counter. However this means that * bytes_pinned does not reflect the bytes that will be pinned once the * delayed refs are flushed, so this counter is inc'ed everytime we call * btrfs_free_extent so it is a realtime count of what will be freed * once the transaction is committed. It will be zero'ed everytime the * transaction commits. * delayed refs are flushed, so this counter is inc'ed every time we * call btrfs_free_extent so it is a realtime count of what will be * freed once the transaction is committed. It will be zero'ed every * time the transaction commits. */ struct percpu_counter total_bytes_pinned; Loading
fs/btrfs/dev-replace.c +1 −1 Original line number Diff line number Diff line Loading @@ -858,7 +858,7 @@ int btrfs_dev_replace_is_ongoing(struct btrfs_dev_replace *dev_replace) * not called and the the filesystem is remounted * in degraded state. This does not stop the * dev_replace procedure. It needs to be canceled * manually if the cancelation is wanted. * manually if the cancellation is wanted. */ break; } Loading
fs/btrfs/disk-io.c +1 −1 Original line number Diff line number Diff line Loading @@ -816,7 +816,7 @@ static void run_one_async_done(struct btrfs_work *work) waitqueue_active(&fs_info->async_submit_wait)) wake_up(&fs_info->async_submit_wait); /* If an error occured we just want to clean up the bio and move on */ /* If an error occurred we just want to clean up the bio and move on */ if (async->error) { async->bio->bi_error = async->error; bio_endio(async->bio); Loading
fs/btrfs/extent-tree.c +1 −1 Original line number Diff line number Diff line Loading @@ -5758,7 +5758,7 @@ int btrfs_delalloc_reserve_metadata(struct inode *inode, u64 num_bytes) /* * This is tricky, but first we need to figure out how much we * free'd from any free-ers that occured during this * free'd from any free-ers that occurred during this * reservation, so we reset ->csum_bytes to the csum_bytes * before we dropped our lock, and then call the free for the * number of bytes that were freed while we were trying our Loading