Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit f7a52a40 authored by David Sterba's avatar David Sterba Committed by Josef Bacik
Browse files

btrfs: remove unused gfp mask parameter from release_extent_buffer callchain



It's unused since 0b32f4bb.

Signed-off-by: default avatarDavid Sterba <dsterba@suse.cz>
Signed-off-by: default avatarJosef Bacik <jbacik@fusionio.com>
parent 34c2b290
Loading
Loading
Loading
Loading
+1 −7
Original line number Diff line number Diff line
@@ -966,14 +966,8 @@ static int btree_releasepage(struct page *page, gfp_t gfp_flags)
{
	if (PageWriteback(page) || PageDirty(page))
		return 0;
	/*
	 * We need to mask out eg. __GFP_HIGHMEM and __GFP_DMA32 as we're doing
	 * slab allocation from alloc_extent_state down the callchain where
	 * it'd hit a BUG_ON as those flags are not allowed.
	 */
	gfp_flags &= ~GFP_SLAB_BUG_MASK;

	return try_release_extent_buffer(page, gfp_flags);
	return try_release_extent_buffer(page);
}

static void btree_invalidatepage(struct page *page, unsigned long offset)
+5 −8
Original line number Diff line number Diff line
@@ -4450,7 +4450,7 @@ static inline void btrfs_release_extent_buffer_rcu(struct rcu_head *head)
}

/* Expects to have eb->eb_lock already held */
static int release_extent_buffer(struct extent_buffer *eb, gfp_t mask)
static int release_extent_buffer(struct extent_buffer *eb)
{
	WARN_ON(atomic_read(&eb->refs) == 0);
	if (atomic_dec_and_test(&eb->refs)) {
@@ -4508,7 +4508,7 @@ void free_extent_buffer(struct extent_buffer *eb)
	 * I know this is terrible, but it's temporary until we stop tracking
	 * the uptodate bits and such for the extent buffers.
	 */
	release_extent_buffer(eb, GFP_ATOMIC);
	release_extent_buffer(eb);
}

void free_extent_buffer_stale(struct extent_buffer *eb)
@@ -4522,7 +4522,7 @@ void free_extent_buffer_stale(struct extent_buffer *eb)
	if (atomic_read(&eb->refs) == 2 && !extent_buffer_under_io(eb) &&
	    test_and_clear_bit(EXTENT_BUFFER_TREE_REF, &eb->bflags))
		atomic_dec(&eb->refs);
	release_extent_buffer(eb, GFP_NOFS);
	release_extent_buffer(eb);
}

void clear_extent_buffer_dirty(struct extent_buffer *eb)
@@ -5042,7 +5042,7 @@ void memmove_extent_buffer(struct extent_buffer *dst, unsigned long dst_offset,
	}
}

int try_release_extent_buffer(struct page *page, gfp_t mask)
int try_release_extent_buffer(struct page *page)
{
	struct extent_buffer *eb;

@@ -5072,9 +5072,6 @@ int try_release_extent_buffer(struct page *page, gfp_t mask)
	}
	spin_unlock(&page->mapping->private_lock);

	if ((mask & GFP_NOFS) == GFP_NOFS)
		mask = GFP_NOFS;

	/*
	 * If tree ref isn't set then we know the ref on this eb is a real ref,
	 * so just return, this page will likely be freed soon anyway.
@@ -5084,5 +5081,5 @@ int try_release_extent_buffer(struct page *page, gfp_t mask)
		return 0;
	}

	return release_extent_buffer(eb, mask);
	return release_extent_buffer(eb);
}
+1 −1
Original line number Diff line number Diff line
@@ -189,7 +189,7 @@ void extent_io_tree_init(struct extent_io_tree *tree,
int try_release_extent_mapping(struct extent_map_tree *map,
			       struct extent_io_tree *tree, struct page *page,
			       gfp_t mask);
int try_release_extent_buffer(struct page *page, gfp_t mask);
int try_release_extent_buffer(struct page *page);
int lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
		     int bits, struct extent_state **cached);