Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit c95d6ed4 authored by Jaegeuk Kim's avatar Jaegeuk Kim
Browse files

Merge remote-tracking branch 'aosp/upstream-f2fs-stable-linux-4.19.y' into android-4.19-stable



This has one fix in advance merged in f2fs-stable.
("xfs: drop I_DIRTY_TIME_EXPIRED")

* aosp/upstream-f2fs-stable-linux-4.19.y:
  writeback: Drop I_DIRTY_TIME_EXPIRE
  writeback: Fix sync livelock due to b_dirty_time processing
  writeback: Avoid skipping inode writeback
  writeback: Protect inode->i_io_list with inode->i_lock
  Revert "writeback: Avoid skipping inode writeback"

Bug: 154542664
Change-Id: I98a6258cb60227e6ca02e57bf7adf28ab7816cbf
Signed-off-by: default avatarJaegeuk Kim <jaegeuk@google.com>
parents 2cadb41e 00f6b03b
Loading
Loading
Loading
Loading
+1 −1
Original line number Original line Diff line number Diff line
@@ -5272,7 +5272,7 @@ static int other_inode_match(struct inode * inode, unsigned long ino,
	    (inode->i_state & I_DIRTY_TIME)) {
	    (inode->i_state & I_DIRTY_TIME)) {
		struct ext4_inode_info	*ei = EXT4_I(inode);
		struct ext4_inode_info	*ei = EXT4_I(inode);


		inode->i_state &= ~(I_DIRTY_TIME | I_DIRTY_TIME_EXPIRED);
		inode->i_state &= ~I_DIRTY_TIME;
		spin_unlock(&inode->i_lock);
		spin_unlock(&inode->i_lock);


		spin_lock(&ei->i_raw_lock);
		spin_lock(&ei->i_raw_lock);
+32 −48
Original line number Original line Diff line number Diff line
@@ -45,7 +45,6 @@ struct wb_completion {
struct wb_writeback_work {
struct wb_writeback_work {
	long nr_pages;
	long nr_pages;
	struct super_block *sb;
	struct super_block *sb;
	unsigned long *older_than_this;
	enum writeback_sync_modes sync_mode;
	enum writeback_sync_modes sync_mode;
	unsigned int tagged_writepages:1;
	unsigned int tagged_writepages:1;
	unsigned int for_kupdate:1;
	unsigned int for_kupdate:1;
@@ -1095,9 +1094,10 @@ void sb_clear_inode_writeback(struct inode *inode)
 * the case then the inode must have been redirtied while it was being written
 * the case then the inode must have been redirtied while it was being written
 * out and we don't reset its dirtied_when.
 * out and we don't reset its dirtied_when.
 */
 */
static void __redirty_tail(struct inode *inode, struct bdi_writeback *wb)
static void redirty_tail_locked(struct inode *inode, struct bdi_writeback *wb)
{
{
	assert_spin_locked(&inode->i_lock);
	assert_spin_locked(&inode->i_lock);

	if (!list_empty(&wb->b_dirty)) {
	if (!list_empty(&wb->b_dirty)) {
		struct inode *tail;
		struct inode *tail;


@@ -1112,7 +1112,7 @@ static void __redirty_tail(struct inode *inode, struct bdi_writeback *wb)
static void redirty_tail(struct inode *inode, struct bdi_writeback *wb)
static void redirty_tail(struct inode *inode, struct bdi_writeback *wb)
{
{
	spin_lock(&inode->i_lock);
	spin_lock(&inode->i_lock);
	__redirty_tail(inode, wb);
	redirty_tail_locked(inode, wb);
	spin_unlock(&inode->i_lock);
	spin_unlock(&inode->i_lock);
}
}


@@ -1152,16 +1152,13 @@ static bool inode_dirtied_after(struct inode *inode, unsigned long t)
#define EXPIRE_DIRTY_ATIME 0x0001
#define EXPIRE_DIRTY_ATIME 0x0001


/*
/*
 * Move expired (dirtied before work->older_than_this) dirty inodes from
 * Move expired (dirtied before dirtied_before) dirty inodes from
 * @delaying_queue to @dispatch_queue.
 * @delaying_queue to @dispatch_queue.
 */
 */
static int move_expired_inodes(struct list_head *delaying_queue,
static int move_expired_inodes(struct list_head *delaying_queue,
			       struct list_head *dispatch_queue,
			       struct list_head *dispatch_queue,
			       int flags,
			       unsigned long dirtied_before)
			       struct wb_writeback_work *work)
{
{
	unsigned long *older_than_this = NULL;
	unsigned long expire_time;
	LIST_HEAD(tmp);
	LIST_HEAD(tmp);
	struct list_head *pos, *node;
	struct list_head *pos, *node;
	struct super_block *sb = NULL;
	struct super_block *sb = NULL;
@@ -1169,22 +1166,13 @@ static int move_expired_inodes(struct list_head *delaying_queue,
	int do_sb_sort = 0;
	int do_sb_sort = 0;
	int moved = 0;
	int moved = 0;


	if ((flags & EXPIRE_DIRTY_ATIME) == 0)
		older_than_this = work->older_than_this;
	else if (!work->for_sync) {
		expire_time = jiffies - (dirtytime_expire_interval * HZ);
		older_than_this = &expire_time;
	}
	while (!list_empty(delaying_queue)) {
	while (!list_empty(delaying_queue)) {
		inode = wb_inode(delaying_queue->prev);
		inode = wb_inode(delaying_queue->prev);
		if (older_than_this &&
		if (inode_dirtied_after(inode, dirtied_before))
		    inode_dirtied_after(inode, *older_than_this))
			break;
			break;
		list_move(&inode->i_io_list, &tmp);
		list_move(&inode->i_io_list, &tmp);
		moved++;
		moved++;
		spin_lock(&inode->i_lock);
		spin_lock(&inode->i_lock);
		if (flags & EXPIRE_DIRTY_ATIME)
			inode->i_state |= I_DIRTY_TIME_EXPIRED;
		inode->i_state |= I_SYNC_QUEUED;
		inode->i_state |= I_SYNC_QUEUED;
		spin_unlock(&inode->i_lock);
		spin_unlock(&inode->i_lock);
		if (sb_is_blkdev_sb(inode->i_sb))
		if (sb_is_blkdev_sb(inode->i_sb))
@@ -1224,18 +1212,22 @@ static int move_expired_inodes(struct list_head *delaying_queue,
 *                                           |
 *                                           |
 *                                           +--> dequeue for IO
 *                                           +--> dequeue for IO
 */
 */
static void queue_io(struct bdi_writeback *wb, struct wb_writeback_work *work)
static void queue_io(struct bdi_writeback *wb, struct wb_writeback_work *work,
		     unsigned long dirtied_before)
{
{
	int moved;
	int moved;
	unsigned long time_expire_jif = dirtied_before;


	assert_spin_locked(&wb->list_lock);
	assert_spin_locked(&wb->list_lock);
	list_splice_init(&wb->b_more_io, &wb->b_io);
	list_splice_init(&wb->b_more_io, &wb->b_io);
	moved = move_expired_inodes(&wb->b_dirty, &wb->b_io, 0, work);
	moved = move_expired_inodes(&wb->b_dirty, &wb->b_io, dirtied_before);
	if (!work->for_sync)
		time_expire_jif = jiffies - dirtytime_expire_interval * HZ;
	moved += move_expired_inodes(&wb->b_dirty_time, &wb->b_io,
	moved += move_expired_inodes(&wb->b_dirty_time, &wb->b_io,
				     EXPIRE_DIRTY_ATIME, work);
				     time_expire_jif);
	if (moved)
	if (moved)
		wb_io_lists_populated(wb);
		wb_io_lists_populated(wb);
	trace_writeback_queue_io(wb, work, moved);
	trace_writeback_queue_io(wb, work, dirtied_before, moved);
}
}


static int write_inode(struct inode *inode, struct writeback_control *wbc)
static int write_inode(struct inode *inode, struct writeback_control *wbc)
@@ -1329,7 +1321,7 @@ static void requeue_inode(struct inode *inode, struct bdi_writeback *wb,
		 * writeback is not making progress due to locked
		 * writeback is not making progress due to locked
		 * buffers. Skip this inode for now.
		 * buffers. Skip this inode for now.
		 */
		 */
		__redirty_tail(inode, wb);
		redirty_tail_locked(inode, wb);
		return;
		return;
	}
	}


@@ -1349,7 +1341,7 @@ static void requeue_inode(struct inode *inode, struct bdi_writeback *wb,
			 * retrying writeback of the dirty page/inode
			 * retrying writeback of the dirty page/inode
			 * that cannot be performed immediately.
			 * that cannot be performed immediately.
			 */
			 */
			__redirty_tail(inode, wb);
			redirty_tail_locked(inode, wb);
		}
		}
	} else if (inode->i_state & I_DIRTY) {
	} else if (inode->i_state & I_DIRTY) {
		/*
		/*
@@ -1357,7 +1349,7 @@ static void requeue_inode(struct inode *inode, struct bdi_writeback *wb,
		 * such as delayed allocation during submission or metadata
		 * such as delayed allocation during submission or metadata
		 * updates after data IO completion.
		 * updates after data IO completion.
		 */
		 */
		__redirty_tail(inode, wb);
		redirty_tail_locked(inode, wb);
	} else if (inode->i_state & I_DIRTY_TIME) {
	} else if (inode->i_state & I_DIRTY_TIME) {
		inode->dirtied_when = jiffies;
		inode->dirtied_when = jiffies;
		inode_io_list_move_locked(inode, wb, &wb->b_dirty_time);
		inode_io_list_move_locked(inode, wb, &wb->b_dirty_time);
@@ -1408,18 +1400,14 @@ __writeback_single_inode(struct inode *inode, struct writeback_control *wbc)
	spin_lock(&inode->i_lock);
	spin_lock(&inode->i_lock);


	dirty = inode->i_state & I_DIRTY;
	dirty = inode->i_state & I_DIRTY;
	if (inode->i_state & I_DIRTY_TIME) {
	if ((inode->i_state & I_DIRTY_TIME) &&
		if ((dirty & I_DIRTY_INODE) ||
	    ((dirty & I_DIRTY_INODE) ||
		    wbc->sync_mode == WB_SYNC_ALL ||
	     wbc->sync_mode == WB_SYNC_ALL || wbc->for_sync ||
		    unlikely(inode->i_state & I_DIRTY_TIME_EXPIRED) ||
	     time_after(jiffies, inode->dirtied_time_when +
		    unlikely(time_after(jiffies,
			dirtytime_expire_interval * HZ))) {
					(inode->dirtied_time_when +
		dirty |= I_DIRTY_TIME;
					 dirtytime_expire_interval * HZ)))) {
			dirty |= I_DIRTY_TIME | I_DIRTY_TIME_EXPIRED;
		trace_writeback_lazytime(inode);
		trace_writeback_lazytime(inode);
	}
	}
	} else
		inode->i_state &= ~I_DIRTY_TIME_EXPIRED;
	inode->i_state &= ~dirty;
	inode->i_state &= ~dirty;


	/*
	/*
@@ -1605,8 +1593,7 @@ static long writeback_sb_inodes(struct super_block *sb,
		 */
		 */
		spin_lock(&inode->i_lock);
		spin_lock(&inode->i_lock);
		if (inode->i_state & (I_NEW | I_FREEING | I_WILL_FREE)) {
		if (inode->i_state & (I_NEW | I_FREEING | I_WILL_FREE)) {
			inode->i_state &= ~I_SYNC_QUEUED;
			redirty_tail_locked(inode, wb);
			__redirty_tail(inode, wb);
			spin_unlock(&inode->i_lock);
			spin_unlock(&inode->i_lock);
			continue;
			continue;
		}
		}
@@ -1748,7 +1735,7 @@ static long writeback_inodes_wb(struct bdi_writeback *wb, long nr_pages,
	blk_start_plug(&plug);
	blk_start_plug(&plug);
	spin_lock(&wb->list_lock);
	spin_lock(&wb->list_lock);
	if (list_empty(&wb->b_io))
	if (list_empty(&wb->b_io))
		queue_io(wb, &work);
		queue_io(wb, &work, jiffies);
	__writeback_inodes_wb(wb, &work);
	__writeback_inodes_wb(wb, &work);
	spin_unlock(&wb->list_lock);
	spin_unlock(&wb->list_lock);
	blk_finish_plug(&plug);
	blk_finish_plug(&plug);
@@ -1768,7 +1755,7 @@ static long writeback_inodes_wb(struct bdi_writeback *wb, long nr_pages,
 * takes longer than a dirty_writeback_interval interval, then leave a
 * takes longer than a dirty_writeback_interval interval, then leave a
 * one-second gap.
 * one-second gap.
 *
 *
 * older_than_this takes precedence over nr_to_write.  So we'll only write back
 * dirtied_before takes precedence over nr_to_write.  So we'll only write back
 * all dirty pages if they are all attached to "old" mappings.
 * all dirty pages if they are all attached to "old" mappings.
 */
 */
static long wb_writeback(struct bdi_writeback *wb,
static long wb_writeback(struct bdi_writeback *wb,
@@ -1776,14 +1763,11 @@ static long wb_writeback(struct bdi_writeback *wb,
{
{
	unsigned long wb_start = jiffies;
	unsigned long wb_start = jiffies;
	long nr_pages = work->nr_pages;
	long nr_pages = work->nr_pages;
	unsigned long oldest_jif;
	unsigned long dirtied_before = jiffies;
	struct inode *inode;
	struct inode *inode;
	long progress;
	long progress;
	struct blk_plug plug;
	struct blk_plug plug;


	oldest_jif = jiffies;
	work->older_than_this = &oldest_jif;

	blk_start_plug(&plug);
	blk_start_plug(&plug);
	spin_lock(&wb->list_lock);
	spin_lock(&wb->list_lock);
	for (;;) {
	for (;;) {
@@ -1817,14 +1801,14 @@ static long wb_writeback(struct bdi_writeback *wb,
		 * safe.
		 * safe.
		 */
		 */
		if (work->for_kupdate) {
		if (work->for_kupdate) {
			oldest_jif = jiffies -
			dirtied_before = jiffies -
				msecs_to_jiffies(dirty_expire_interval * 10);
				msecs_to_jiffies(dirty_expire_interval * 10);
		} else if (work->for_background)
		} else if (work->for_background)
			oldest_jif = jiffies;
			dirtied_before = jiffies;


		trace_writeback_start(wb, work);
		trace_writeback_start(wb, work);
		if (list_empty(&wb->b_io))
		if (list_empty(&wb->b_io))
			queue_io(wb, work);
			queue_io(wb, work, dirtied_before);
		if (work->sb)
		if (work->sb)
			progress = writeback_sb_inodes(work->sb, wb, work);
			progress = writeback_sb_inodes(work->sb, wb, work);
		else
		else
+2 −2
Original line number Original line Diff line number Diff line
@@ -97,9 +97,9 @@ xfs_trans_log_inode(
	 * to log the timestamps, or will clear already cleared fields in the
	 * to log the timestamps, or will clear already cleared fields in the
	 * worst case.
	 * worst case.
	 */
	 */
	if (inode->i_state & (I_DIRTY_TIME | I_DIRTY_TIME_EXPIRED)) {
	if (inode->i_state & I_DIRTY_TIME) {
		spin_lock(&inode->i_lock);
		spin_lock(&inode->i_lock);
		inode->i_state &= ~(I_DIRTY_TIME | I_DIRTY_TIME_EXPIRED);
		inode->i_state &= ~I_DIRTY_TIME;
		spin_unlock(&inode->i_lock);
		spin_unlock(&inode->i_lock);
	}
	}


+1 −2
Original line number Original line Diff line number Diff line
@@ -2162,11 +2162,10 @@ static inline void init_sync_kiocb(struct kiocb *kiocb, struct file *filp)
#define I_DIO_WAKEUP		(1 << __I_DIO_WAKEUP)
#define I_DIO_WAKEUP		(1 << __I_DIO_WAKEUP)
#define I_LINKABLE		(1 << 10)
#define I_LINKABLE		(1 << 10)
#define I_DIRTY_TIME		(1 << 11)
#define I_DIRTY_TIME		(1 << 11)
#define I_DIRTY_TIME_EXPIRED	(1 << 12)
#define I_WB_SWITCH		(1 << 13)
#define I_WB_SWITCH		(1 << 13)
#define I_OVL_INUSE		(1 << 14)
#define I_OVL_INUSE		(1 << 14)
#define I_CREATING		(1 << 15)
#define I_CREATING		(1 << 15)
#define I_SYNC_QUEUED		(1 << 16)
#define I_SYNC_QUEUED		(1 << 17)


#define I_DIRTY_INODE (I_DIRTY_SYNC | I_DIRTY_DATASYNC)
#define I_DIRTY_INODE (I_DIRTY_SYNC | I_DIRTY_DATASYNC)
#define I_DIRTY (I_DIRTY_INODE | I_DIRTY_PAGES)
#define I_DIRTY (I_DIRTY_INODE | I_DIRTY_PAGES)
+6 −8
Original line number Original line Diff line number Diff line
@@ -20,7 +20,6 @@
		{I_CLEAR,		"I_CLEAR"},		\
		{I_CLEAR,		"I_CLEAR"},		\
		{I_SYNC,		"I_SYNC"},		\
		{I_SYNC,		"I_SYNC"},		\
		{I_DIRTY_TIME,		"I_DIRTY_TIME"},	\
		{I_DIRTY_TIME,		"I_DIRTY_TIME"},	\
		{I_DIRTY_TIME_EXPIRED,	"I_DIRTY_TIME_EXPIRED"}, \
		{I_REFERENCED,		"I_REFERENCED"}		\
		{I_REFERENCED,		"I_REFERENCED"}		\
	)
	)


@@ -360,8 +359,9 @@ DEFINE_WBC_EVENT(wbc_writepage);
TRACE_EVENT(writeback_queue_io,
TRACE_EVENT(writeback_queue_io,
	TP_PROTO(struct bdi_writeback *wb,
	TP_PROTO(struct bdi_writeback *wb,
		 struct wb_writeback_work *work,
		 struct wb_writeback_work *work,
		 unsigned long dirtied_before,
		 int moved),
		 int moved),
	TP_ARGS(wb, work, moved),
	TP_ARGS(wb, work, dirtied_before, moved),
	TP_STRUCT__entry(
	TP_STRUCT__entry(
		__array(char,		name, 32)
		__array(char,		name, 32)
		__field(unsigned long,	older)
		__field(unsigned long,	older)
@@ -371,19 +371,17 @@ TRACE_EVENT(writeback_queue_io,
		__field(unsigned int,	cgroup_ino)
		__field(unsigned int,	cgroup_ino)
	),
	),
	TP_fast_assign(
	TP_fast_assign(
		unsigned long *older_than_this = work->older_than_this;
		strncpy(__entry->name, dev_name(wb->bdi->dev), 32);
		strncpy(__entry->name, dev_name(wb->bdi->dev), 32);
		__entry->older	= older_than_this ?  *older_than_this : 0;
		__entry->older	= dirtied_before;
		__entry->age	= older_than_this ?
		__entry->age	= (jiffies - dirtied_before) * 1000 / HZ;
				  (jiffies - *older_than_this) * 1000 / HZ : -1;
		__entry->moved	= moved;
		__entry->moved	= moved;
		__entry->reason	= work->reason;
		__entry->reason	= work->reason;
		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
		__entry->cgroup_ino	= __trace_wb_assign_cgroup(wb);
	),
	),
	TP_printk("bdi %s: older=%lu age=%ld enqueue=%d reason=%s cgroup_ino=%u",
	TP_printk("bdi %s: older=%lu age=%ld enqueue=%d reason=%s cgroup_ino=%u",
		__entry->name,
		__entry->name,
		__entry->older,	/* older_than_this in jiffies */
		__entry->older,	/* dirtied_before in jiffies */
		__entry->age,	/* older_than_this in relative milliseconds */
		__entry->age,	/* dirtied_before in relative milliseconds */
		__entry->moved,
		__entry->moved,
		__print_symbolic(__entry->reason, WB_WORK_REASON),
		__print_symbolic(__entry->reason, WB_WORK_REASON),
		__entry->cgroup_ino
		__entry->cgroup_ino