2017-08-05 03:04:55

by Wang Shilong

[permalink] [raw]
Subject: [PATCH v2] ext4: reduce lock contention in __ext4_new_inode

While running number of creating file threads concurrently,
we found heavy lock contention on group spinlock:

FUNC TOTAL_TIME(us) COUNT AVG(us)
ext4_create 1707443399 1440000 1185.72
_raw_spin_lock 1317641501 180899929 7.28
jbd2__journal_start 287821030 1453950 197.96
jbd2_journal_get_write_access 33441470 73077185 0.46
ext4_add_nondir 29435963 1440000 20.44
ext4_add_entry 26015166 1440049 18.07
ext4_dx_add_entry 25729337 1432814 17.96
ext4_mark_inode_dirty 12302433 5774407 2.13

most of cpu time blames to _raw_spin_lock, here is some testing
numbers with/without patch.

Test environment:
Server : SuperMicro Sever (2 x E5-2690 [email protected], 128GB 2133MHz
DDR4 Memory, 8GbFC)
Storage : 2 x RAID1 (DDN SFA7700X, 4 x Toshiba PX02SMU020 200GB
Read Intensive SSD)

format command:
mkfs.ext4 -J size=4096

test command:
mpirun -np 48 mdtest -n 30000 -d /ext4/mdtest.out -F -C \
-r -i 5 -v -p 10 -u

Kernel version: 4.13.0-rc3

Test 1,440,000 files with 48 directories by 48 processes:

Without patch:

File Creation File removal
79,033 289,569 ops/per second
81,463 285,359
79,875 288,475
79,917 284,624
79,420 290,91

ith patch:
File Creation File removal
302,600 312,813 ops/per second
295,644 316,557
288,125 306,961
302,960 310,517
295,175 311,927

Now create and removal performaces are similar, and creation
performaces are improved more than 3x with large journal size.
In default journal size, performances are improved by 50%.

Tested-by: Shuichi Ihara <[email protected]>
Signed-off-by: Wang Shilong <[email protected]>
---
v1->v2: use ext4_fs_is_busy() helper.
---
fs/ext4/ialloc.c | 7 ++++++-
1 file changed, 6 insertions(+), 1 deletion(-)

diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c
index 507bfb3..19323ea 100644
--- a/fs/ext4/ialloc.c
+++ b/fs/ext4/ialloc.c
@@ -957,8 +957,13 @@ struct inode *__ext4_new_inode(handle_t *handle, struct inode *dir,
if (!ret2)
goto got; /* we grabbed the inode! */
next_inode:
- if (ino < EXT4_INODES_PER_GROUP(sb))
+ if (ino < EXT4_INODES_PER_GROUP(sb)) {
+ /* Lock contention, relax a bit */
+ if (ext4_fs_is_busy(sbi))
+ schedule_timeout_uninterruptible(
+ msecs_to_jiffies(1));
goto repeat_in_this_group;
+ }
next_group:
if (++group == ngroups)
group = 0;
--
2.9.3


2017-08-05 17:03:57

by Theodore Ts'o

[permalink] [raw]
Subject: Re: [PATCH v2] ext4: reduce lock contention in __ext4_new_inode

On Sat, Aug 05, 2017 at 11:04:36AM +0800, Wang Shilong wrote:
> diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c
> index 507bfb3..19323ea 100644
> --- a/fs/ext4/ialloc.c
> +++ b/fs/ext4/ialloc.c
> @@ -957,8 +957,13 @@ struct inode *__ext4_new_inode(handle_t *handle, struct inode *dir,
> if (!ret2)
> goto got; /* we grabbed the inode! */
> next_inode:
> - if (ino < EXT4_INODES_PER_GROUP(sb))
> + if (ino < EXT4_INODES_PER_GROUP(sb)) {
> + /* Lock contention, relax a bit */
> + if (ext4_fs_is_busy(sbi))
> + schedule_timeout_uninterruptible(
> + msecs_to_jiffies(1));
> goto repeat_in_this_group;
> + }
> next_group:
> if (++group == ngroups)
> group = 0;

We should probably ne not even doing the lock contention in the case
where the reason why we've jumped to next_inode is because we failed
the recently_deleted() test. But that can be fixed by changing the
"goto next_inode" in the recently_deleted() codepath with:

if (ino < EXT4_INODES_PER_GROUP(sb))
goto repeat_in_this_group;

Also while I agree that it's better to use ext4_fs_is_busy(), the
exact details of when we will sleep for a second are different. So it
would be good for you to rerun your benchmarks; since the numbers in
your v1 and v2 patch were the same, it's not clear to me that you did
rerun them. Can you confirm one way or another? And rerun them for
the v3 version of the patch?

Many thanks,

- Ted

2017-08-06 00:15:51

by Wang Shilong

[permalink] [raw]
Subject: Re: [PATCH v2] ext4: reduce lock contention in __ext4_new_inode

On Sun, Aug 6, 2017 at 1:03 AM, Theodore Ts'o <[email protected]> wrote:
> On Sat, Aug 05, 2017 at 11:04:36AM +0800, Wang Shilong wrote:
>> diff --git a/fs/ext4/ialloc.c b/fs/ext4/ialloc.c
>> index 507bfb3..19323ea 100644
>> --- a/fs/ext4/ialloc.c
>> +++ b/fs/ext4/ialloc.c
>> @@ -957,8 +957,13 @@ struct inode *__ext4_new_inode(handle_t *handle, struct inode *dir,
>> if (!ret2)
>> goto got; /* we grabbed the inode! */
>> next_inode:
>> - if (ino < EXT4_INODES_PER_GROUP(sb))
>> + if (ino < EXT4_INODES_PER_GROUP(sb)) {
>> + /* Lock contention, relax a bit */
>> + if (ext4_fs_is_busy(sbi))
>> + schedule_timeout_uninterruptible(
>> + msecs_to_jiffies(1));
>> goto repeat_in_this_group;
>> + }
>> next_group:
>> if (++group == ngroups)
>> group = 0;
>
> We should probably ne not even doing the lock contention in the case
> where the reason why we've jumped to next_inode is because we failed
> the recently_deleted() test. But that can be fixed by changing the
> "goto next_inode" in the recently_deleted() codepath with:
>
> if (ino < EXT4_INODES_PER_GROUP(sb))
> goto repeat_in_this_group;
>

Yup, you are right, i thought about that in the first patch, but missed
it when v2.


> Also while I agree that it's better to use ext4_fs_is_busy(), the
> exact details of when we will sleep for a second are different. So it
> would be good for you to rerun your benchmarks; since the numbers in
> your v1 and v2 patch were the same, it's not clear to me that you did
> rerun them. Can you confirm one way or another? And rerun them for
> the v3 version of the patch?

We indeed should rerun benchmark, thanks for your timely feedback, will
rebenchmark as you suggested.


>
> Many thanks,
>
> - Ted