2019-03-15 11:12:20

by Luis Henriques

[permalink] [raw]
Subject: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

I'm occasionally seeing a kmemleak warning in xfstest generic/013:

unreferenced object 0xffff8881fccca940 (size 32):
comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
hex dump (first 32 bytes):
01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
backtrace:
[<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
[<0000000021a00533>] rebuild_snap_realms+0x27/0x90
[<00000000ac538600>] rebuild_snap_realms+0x42/0x90
[<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
[<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
[<00000000fc287b83>] dispatch+0x362/0x176c
[<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
[<000000004168e3a9>] process_one_work+0x1d4/0x400
[<000000002188e9e7>] worker_thread+0x2d/0x3c0
[<00000000b593e4b3>] kthread+0x112/0x130
[<00000000a8587dca>] ret_from_fork+0x35/0x40
[<00000000ba1c9c1d>] 0xffffffffffffffff

It looks like it is possible that we miss a flush_ack from the MDS when,
for example, umounting the filesystem. In that case, we can simply drop
the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().

Link: https://tracker.ceph.com/issues/38224
Cc: [email protected]
Signed-off-by: Luis Henriques <[email protected]>
---
fs/ceph/caps.c | 7 +++++++
1 file changed, 7 insertions(+)

diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
index 36a8dc699448..208f4dc6f574 100644
--- a/fs/ceph/caps.c
+++ b/fs/ceph/caps.c
@@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
static void drop_inode_snap_realm(struct ceph_inode_info *ci)
{
struct ceph_snap_realm *realm = ci->i_snap_realm;
+
spin_lock(&realm->inodes_with_caps_lock);
list_del_init(&ci->i_snap_realm_item);
ci->i_snap_realm_counter++;
@@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
spin_unlock(&realm->inodes_with_caps_lock);
ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
realm);
+ /*
+ * ci->i_head_snapc should be NULL, but we may still be waiting for a
+ * flush_ack from the MDS. In that case, we still hold a ref for the
+ * ceph_snap_context and we need to drop it.
+ */
+ ceph_put_snap_context(ci->i_head_snapc);
}

/*


2019-03-18 03:27:33

by Yan, Zheng

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>
> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>
> unreferenced object 0xffff8881fccca940 (size 32):
> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
> hex dump (first 32 bytes):
> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
> backtrace:
> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
> [<00000000fc287b83>] dispatch+0x362/0x176c
> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
> [<000000004168e3a9>] process_one_work+0x1d4/0x400
> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
> [<00000000b593e4b3>] kthread+0x112/0x130
> [<00000000a8587dca>] ret_from_fork+0x35/0x40
> [<00000000ba1c9c1d>] 0xffffffffffffffff
>
> It looks like it is possible that we miss a flush_ack from the MDS when,
> for example, umounting the filesystem. In that case, we can simply drop
> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>
> Link: https://tracker.ceph.com/issues/38224
> Cc: [email protected]
> Signed-off-by: Luis Henriques <[email protected]>
> ---
> fs/ceph/caps.c | 7 +++++++
> 1 file changed, 7 insertions(+)
>
> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
> index 36a8dc699448..208f4dc6f574 100644
> --- a/fs/ceph/caps.c
> +++ b/fs/ceph/caps.c
> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> {
> struct ceph_snap_realm *realm = ci->i_snap_realm;
> +
> spin_lock(&realm->inodes_with_caps_lock);
> list_del_init(&ci->i_snap_realm_item);
> ci->i_snap_realm_counter++;
> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> spin_unlock(&realm->inodes_with_caps_lock);
> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
> realm);
> + /*
> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
> + * flush_ack from the MDS. In that case, we still hold a ref for the
> + * ceph_snap_context and we need to drop it.
> + */
> + ceph_put_snap_context(ci->i_head_snapc);
> }
>
> /*

This does not seem right. i_head_snapc is cleared when
(ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
Did you see 'reconnect denied' during the test? If you did, the fix
should be in iterate_session_caps()

2019-03-18 10:34:45

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

"Yan, Zheng" <[email protected]> writes:

> On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>>
>> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>>
>> unreferenced object 0xffff8881fccca940 (size 32):
>> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>> hex dump (first 32 bytes):
>> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>> backtrace:
>> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>> [<00000000fc287b83>] dispatch+0x362/0x176c
>> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>> [<00000000b593e4b3>] kthread+0x112/0x130
>> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>> [<00000000ba1c9c1d>] 0xffffffffffffffff
>>
>> It looks like it is possible that we miss a flush_ack from the MDS when,
>> for example, umounting the filesystem. In that case, we can simply drop
>> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>>
>> Link: https://tracker.ceph.com/issues/38224
>> Cc: [email protected]
>> Signed-off-by: Luis Henriques <[email protected]>
>> ---
>> fs/ceph/caps.c | 7 +++++++
>> 1 file changed, 7 insertions(+)
>>
>> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>> index 36a8dc699448..208f4dc6f574 100644
>> --- a/fs/ceph/caps.c
>> +++ b/fs/ceph/caps.c
>> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> {
>> struct ceph_snap_realm *realm = ci->i_snap_realm;
>> +
>> spin_lock(&realm->inodes_with_caps_lock);
>> list_del_init(&ci->i_snap_realm_item);
>> ci->i_snap_realm_counter++;
>> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> spin_unlock(&realm->inodes_with_caps_lock);
>> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>> realm);
>> + /*
>> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>> + * flush_ack from the MDS. In that case, we still hold a ref for the
>> + * ceph_snap_context and we need to drop it.
>> + */
>> + ceph_put_snap_context(ci->i_head_snapc);
>> }
>>
>> /*
>
> This does not seem right. i_head_snapc is cleared when
> (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
> ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
> Did you see 'reconnect denied' during the test? If you did, the fix
> should be in iterate_session_caps()
>

No, I didn't saw any 'reconnect denied' in the test. The test actually
seems to execute fine, except from the memory leak.

It's very difficult to reproduce this issue, but last time I managed to
get this memory leak to trigger I actually had some debugging code in
drop_inode_snap_realm, something like:

if (ci->i_head_snapc)
printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);

This printk was only executed when the bug triggered (during a
filesystem umount) and the address shown was the same as in the kmemleak
warning.

After spending some time looking, I assumed this to be a missing call to
handle_cap_flush_ack, which would do the i_head_snapc cleanup.

Cheers,
--
Luis

2019-03-18 11:43:59

by Yan, Zheng

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>
> "Yan, Zheng" <[email protected]> writes:
>
> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
> >>
> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
> >>
> >> unreferenced object 0xffff8881fccca940 (size 32):
> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
> >> hex dump (first 32 bytes):
> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
> >> backtrace:
> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
> >> [<00000000fc287b83>] dispatch+0x362/0x176c
> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
> >> [<00000000b593e4b3>] kthread+0x112/0x130
> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
> >>
> >> It looks like it is possible that we miss a flush_ack from the MDS when,
> >> for example, umounting the filesystem. In that case, we can simply drop
> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
> >>
> >> Link: https://tracker.ceph.com/issues/38224
> >> Cc: [email protected]
> >> Signed-off-by: Luis Henriques <[email protected]>
> >> ---
> >> fs/ceph/caps.c | 7 +++++++
> >> 1 file changed, 7 insertions(+)
> >>
> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
> >> index 36a8dc699448..208f4dc6f574 100644
> >> --- a/fs/ceph/caps.c
> >> +++ b/fs/ceph/caps.c
> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >> {
> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
> >> +
> >> spin_lock(&realm->inodes_with_caps_lock);
> >> list_del_init(&ci->i_snap_realm_item);
> >> ci->i_snap_realm_counter++;
> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >> spin_unlock(&realm->inodes_with_caps_lock);
> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
> >> realm);
> >> + /*
> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
> >> + * ceph_snap_context and we need to drop it.
> >> + */
> >> + ceph_put_snap_context(ci->i_head_snapc);
> >> }
> >>
> >> /*
> >
> > This does not seem right. i_head_snapc is cleared when
> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
> > Did you see 'reconnect denied' during the test? If you did, the fix
> > should be in iterate_session_caps()
> >
>
> No, I didn't saw any 'reconnect denied' in the test. The test actually
> seems to execute fine, except from the memory leak.
>
> It's very difficult to reproduce this issue, but last time I managed to
> get this memory leak to trigger I actually had some debugging code in
> drop_inode_snap_realm, something like:
>
> if (ci->i_head_snapc)
> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);

please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
i_flushing_caps. and try reproducing it again.


>
> This printk was only executed when the bug triggered (during a
> filesystem umount) and the address shown was the same as in the kmemleak
> warning.
>
> After spending some time looking, I assumed this to be a missing call to
> handle_cap_flush_ack, which would do the i_head_snapc cleanup.
>
> Cheers,
> --
> Luis

2019-03-18 16:23:46

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

"Yan, Zheng" <[email protected]> writes:

> On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>>
>> "Yan, Zheng" <[email protected]> writes:
>>
>> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>> >>
>> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>> >>
>> >> unreferenced object 0xffff8881fccca940 (size 32):
>> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>> >> hex dump (first 32 bytes):
>> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>> >> backtrace:
>> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>> >> [<00000000fc287b83>] dispatch+0x362/0x176c
>> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>> >> [<00000000b593e4b3>] kthread+0x112/0x130
>> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
>> >>
>> >> It looks like it is possible that we miss a flush_ack from the MDS when,
>> >> for example, umounting the filesystem. In that case, we can simply drop
>> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>> >>
>> >> Link: https://tracker.ceph.com/issues/38224
>> >> Cc: [email protected]
>> >> Signed-off-by: Luis Henriques <[email protected]>
>> >> ---
>> >> fs/ceph/caps.c | 7 +++++++
>> >> 1 file changed, 7 insertions(+)
>> >>
>> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>> >> index 36a8dc699448..208f4dc6f574 100644
>> >> --- a/fs/ceph/caps.c
>> >> +++ b/fs/ceph/caps.c
>> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >> {
>> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
>> >> +
>> >> spin_lock(&realm->inodes_with_caps_lock);
>> >> list_del_init(&ci->i_snap_realm_item);
>> >> ci->i_snap_realm_counter++;
>> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >> spin_unlock(&realm->inodes_with_caps_lock);
>> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>> >> realm);
>> >> + /*
>> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
>> >> + * ceph_snap_context and we need to drop it.
>> >> + */
>> >> + ceph_put_snap_context(ci->i_head_snapc);
>> >> }
>> >>
>> >> /*
>> >
>> > This does not seem right. i_head_snapc is cleared when
>> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
>> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
>> > Did you see 'reconnect denied' during the test? If you did, the fix
>> > should be in iterate_session_caps()
>> >
>>
>> No, I didn't saw any 'reconnect denied' in the test. The test actually
>> seems to execute fine, except from the memory leak.
>>
>> It's very difficult to reproduce this issue, but last time I managed to
>> get this memory leak to trigger I actually had some debugging code in
>> drop_inode_snap_realm, something like:
>>
>> if (ci->i_head_snapc)
>> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
>
> please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
> i_flushing_caps. and try reproducing it again.
>

Ok, it took me a few hours, but I managed to reproduce the bug, with
those extra printks. All those values are set to 0 when the bug
triggers (and i_head_snapc != NULL).

Cheers,
--
Luis


>
>>
>> This printk was only executed when the bug triggered (during a
>> filesystem umount) and the address shown was the same as in the kmemleak
>> warning.
>>
>> After spending some time looking, I assumed this to be a missing call to
>> handle_cap_flush_ack, which would do the i_head_snapc cleanup.
>>
>> Cheers,
>> --
>> Luis
>

2019-03-19 03:17:44

by Yan, Zheng

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
>
> "Yan, Zheng" <[email protected]> writes:
>
> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
> >>
> >> "Yan, Zheng" <[email protected]> writes:
> >>
> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
> >> >>
> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
> >> >>
> >> >> unreferenced object 0xffff8881fccca940 (size 32):
> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
> >> >> hex dump (first 32 bytes):
> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
> >> >> backtrace:
> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
> >> >>
> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
> >> >> for example, umounting the filesystem. In that case, we can simply drop
> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
> >> >>
> >> >> Link: https://tracker.ceph.com/issues/38224
> >> >> Cc: [email protected]
> >> >> Signed-off-by: Luis Henriques <[email protected]>
> >> >> ---
> >> >> fs/ceph/caps.c | 7 +++++++
> >> >> 1 file changed, 7 insertions(+)
> >> >>
> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
> >> >> index 36a8dc699448..208f4dc6f574 100644
> >> >> --- a/fs/ceph/caps.c
> >> >> +++ b/fs/ceph/caps.c
> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >> >> {
> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
> >> >> +
> >> >> spin_lock(&realm->inodes_with_caps_lock);
> >> >> list_del_init(&ci->i_snap_realm_item);
> >> >> ci->i_snap_realm_counter++;
> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >> >> spin_unlock(&realm->inodes_with_caps_lock);
> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
> >> >> realm);
> >> >> + /*
> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
> >> >> + * ceph_snap_context and we need to drop it.
> >> >> + */
> >> >> + ceph_put_snap_context(ci->i_head_snapc);
> >> >> }
> >> >>
> >> >> /*
> >> >
> >> > This does not seem right. i_head_snapc is cleared when
> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
> >> > Did you see 'reconnect denied' during the test? If you did, the fix
> >> > should be in iterate_session_caps()
> >> >
> >>
> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
> >> seems to execute fine, except from the memory leak.
> >>
> >> It's very difficult to reproduce this issue, but last time I managed to
> >> get this memory leak to trigger I actually had some debugging code in
> >> drop_inode_snap_realm, something like:
> >>
> >> if (ci->i_head_snapc)
> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
> >
> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
> > i_flushing_caps. and try reproducing it again.
> >
>
> Ok, it took me a few hours, but I managed to reproduce the bug, with
> those extra printks. All those values are set to 0 when the bug
> triggers (and i_head_snapc != NULL).
>

Thanks, which test triggers this bug?

I searched that code, found we may fail to cleanup i_head_snap in two
places. One is in ceph_queue_cap_snap, Another is in
remove_session_caps_cb().

> Cheers,
> --
> Luis
>
>
> >
> >>
> >> This printk was only executed when the bug triggered (during a
> >> filesystem umount) and the address shown was the same as in the kmemleak
> >> warning.
> >>
> >> After spending some time looking, I assumed this to be a missing call to
> >> handle_cap_flush_ack, which would do the i_head_snapc cleanup.
> >>
> >> Cheers,
> >> --
> >> Luis
> >

2019-03-19 09:41:48

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

"Yan, Zheng" <[email protected]> writes:

> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
>>
>> "Yan, Zheng" <[email protected]> writes:
>>
>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>> >>
>> >> "Yan, Zheng" <[email protected]> writes:
>> >>
>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>> >> >>
>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>> >> >>
>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>> >> >> hex dump (first 32 bytes):
>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>> >> >> backtrace:
>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
>> >> >>
>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
>> >> >> for example, umounting the filesystem. In that case, we can simply drop
>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>> >> >>
>> >> >> Link: https://tracker.ceph.com/issues/38224
>> >> >> Cc: [email protected]
>> >> >> Signed-off-by: Luis Henriques <[email protected]>
>> >> >> ---
>> >> >> fs/ceph/caps.c | 7 +++++++
>> >> >> 1 file changed, 7 insertions(+)
>> >> >>
>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>> >> >> index 36a8dc699448..208f4dc6f574 100644
>> >> >> --- a/fs/ceph/caps.c
>> >> >> +++ b/fs/ceph/caps.c
>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >> >> {
>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
>> >> >> +
>> >> >> spin_lock(&realm->inodes_with_caps_lock);
>> >> >> list_del_init(&ci->i_snap_realm_item);
>> >> >> ci->i_snap_realm_counter++;
>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>> >> >> realm);
>> >> >> + /*
>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
>> >> >> + * ceph_snap_context and we need to drop it.
>> >> >> + */
>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
>> >> >> }
>> >> >>
>> >> >> /*
>> >> >
>> >> > This does not seem right. i_head_snapc is cleared when
>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
>> >> > should be in iterate_session_caps()
>> >> >
>> >>
>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
>> >> seems to execute fine, except from the memory leak.
>> >>
>> >> It's very difficult to reproduce this issue, but last time I managed to
>> >> get this memory leak to trigger I actually had some debugging code in
>> >> drop_inode_snap_realm, something like:
>> >>
>> >> if (ci->i_head_snapc)
>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
>> >
>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
>> > i_flushing_caps. and try reproducing it again.
>> >
>>
>> Ok, it took me a few hours, but I managed to reproduce the bug, with
>> those extra printks. All those values are set to 0 when the bug
>> triggers (and i_head_snapc != NULL).
>>
>
> Thanks, which test triggers this bug?

That's generic/013. It usually triggers after a few hours of running it
in a loop (I'm using a vstart cluster for that).

>
> I searched that code, found we may fail to cleanup i_head_snap in two
> places. One is in ceph_queue_cap_snap, Another is in
> remove_session_caps_cb().

Ah, great! I spent a lot of time looking but I couldn't really find it.
My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
and that the corresponding ceph_put_snap_context would be done in
handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
MDS.

Cheers,
--
Luis


>
>> Cheers,
>> --
>> Luis
>>
>>
>> >
>> >>
>> >> This printk was only executed when the bug triggered (during a
>> >> filesystem umount) and the address shown was the same as in the kmemleak
>> >> warning.
>> >>
>> >> After spending some time looking, I assumed this to be a missing call to
>> >> handle_cap_flush_ack, which would do the i_head_snapc cleanup.
>> >>
>> >> Cheers,
>> >> --
>> >> Luis
>> >
>

2019-03-22 10:06:08

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

Luis Henriques <[email protected]> writes:

> "Yan, Zheng" <[email protected]> writes:
>
>> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
>>>
>>> "Yan, Zheng" <[email protected]> writes:
>>>
>>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>>> >>
>>> >> "Yan, Zheng" <[email protected]> writes:
>>> >>
>>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>>> >> >>
>>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>>> >> >>
>>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
>>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>>> >> >> hex dump (first 32 bytes):
>>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>>> >> >> backtrace:
>>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
>>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
>>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
>>> >> >>
>>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
>>> >> >> for example, umounting the filesystem. In that case, we can simply drop
>>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>>> >> >>
>>> >> >> Link: https://tracker.ceph.com/issues/38224
>>> >> >> Cc: [email protected]
>>> >> >> Signed-off-by: Luis Henriques <[email protected]>
>>> >> >> ---
>>> >> >> fs/ceph/caps.c | 7 +++++++
>>> >> >> 1 file changed, 7 insertions(+)
>>> >> >>
>>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>>> >> >> index 36a8dc699448..208f4dc6f574 100644
>>> >> >> --- a/fs/ceph/caps.c
>>> >> >> +++ b/fs/ceph/caps.c
>>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>>> >> >> {
>>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
>>> >> >> +
>>> >> >> spin_lock(&realm->inodes_with_caps_lock);
>>> >> >> list_del_init(&ci->i_snap_realm_item);
>>> >> >> ci->i_snap_realm_counter++;
>>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
>>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>>> >> >> realm);
>>> >> >> + /*
>>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
>>> >> >> + * ceph_snap_context and we need to drop it.
>>> >> >> + */
>>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
>>> >> >> }
>>> >> >>
>>> >> >> /*
>>> >> >
>>> >> > This does not seem right. i_head_snapc is cleared when
>>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
>>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
>>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
>>> >> > should be in iterate_session_caps()
>>> >> >
>>> >>
>>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
>>> >> seems to execute fine, except from the memory leak.
>>> >>
>>> >> It's very difficult to reproduce this issue, but last time I managed to
>>> >> get this memory leak to trigger I actually had some debugging code in
>>> >> drop_inode_snap_realm, something like:
>>> >>
>>> >> if (ci->i_head_snapc)
>>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
>>> >
>>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
>>> > i_flushing_caps. and try reproducing it again.
>>> >
>>>
>>> Ok, it took me a few hours, but I managed to reproduce the bug, with
>>> those extra printks. All those values are set to 0 when the bug
>>> triggers (and i_head_snapc != NULL).
>>>
>>
>> Thanks, which test triggers this bug?
>
> That's generic/013. It usually triggers after a few hours of running it
> in a loop (I'm using a vstart cluster for that).
>
>>
>> I searched that code, found we may fail to cleanup i_head_snap in two
>> places. One is in ceph_queue_cap_snap, Another is in
>> remove_session_caps_cb().
>
> Ah, great! I spent a lot of time looking but I couldn't really find it.
> My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
> and that the corresponding ceph_put_snap_context would be done in
> handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
> MDS.

Something that I didn't said explicitly is that I *know* that the
unbalanced ceph_get_snap_context() is the one in function
ceph_queue_cap_snap(). I know this for sure because I've managed to
reproduce the issue several times with an instrumented kernel (with
tracepoints) that allowed me to keep track of all ceph_snap_context
operations (create, get, put, delete).

Cheers,
--
Luis

2019-04-03 02:56:46

by Yan, Zheng

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

On Fri, Mar 22, 2019 at 6:04 PM Luis Henriques <[email protected]> wrote:
>
> Luis Henriques <[email protected]> writes:
>
> > "Yan, Zheng" <[email protected]> writes:
> >
> >> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
> >>>
> >>> "Yan, Zheng" <[email protected]> writes:
> >>>
> >>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
> >>> >>
> >>> >> "Yan, Zheng" <[email protected]> writes:
> >>> >>
> >>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
> >>> >> >>
> >>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
> >>> >> >>
> >>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
> >>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
> >>> >> >> hex dump (first 32 bytes):
> >>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
> >>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
> >>> >> >> backtrace:
> >>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
> >>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
> >>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
> >>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
> >>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
> >>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
> >>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
> >>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
> >>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
> >>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
> >>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
> >>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
> >>> >> >>
> >>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
> >>> >> >> for example, umounting the filesystem. In that case, we can simply drop
> >>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
> >>> >> >>
> >>> >> >> Link: https://tracker.ceph.com/issues/38224
> >>> >> >> Cc: [email protected]
> >>> >> >> Signed-off-by: Luis Henriques <[email protected]>
> >>> >> >> ---
> >>> >> >> fs/ceph/caps.c | 7 +++++++
> >>> >> >> 1 file changed, 7 insertions(+)
> >>> >> >>
> >>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
> >>> >> >> index 36a8dc699448..208f4dc6f574 100644
> >>> >> >> --- a/fs/ceph/caps.c
> >>> >> >> +++ b/fs/ceph/caps.c
> >>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
> >>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >>> >> >> {
> >>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
> >>> >> >> +
> >>> >> >> spin_lock(&realm->inodes_with_caps_lock);
> >>> >> >> list_del_init(&ci->i_snap_realm_item);
> >>> >> >> ci->i_snap_realm_counter++;
> >>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
> >>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
> >>> >> >> realm);
> >>> >> >> + /*
> >>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
> >>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
> >>> >> >> + * ceph_snap_context and we need to drop it.
> >>> >> >> + */
> >>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
> >>> >> >> }
> >>> >> >>
> >>> >> >> /*
> >>> >> >
> >>> >> > This does not seem right. i_head_snapc is cleared when
> >>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
> >>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
> >>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
> >>> >> > should be in iterate_session_caps()
> >>> >> >
> >>> >>
> >>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
> >>> >> seems to execute fine, except from the memory leak.
> >>> >>
> >>> >> It's very difficult to reproduce this issue, but last time I managed to
> >>> >> get this memory leak to trigger I actually had some debugging code in
> >>> >> drop_inode_snap_realm, something like:
> >>> >>
> >>> >> if (ci->i_head_snapc)
> >>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
> >>> >
> >>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
> >>> > i_flushing_caps. and try reproducing it again.
> >>> >
> >>>
> >>> Ok, it took me a few hours, but I managed to reproduce the bug, with
> >>> those extra printks. All those values are set to 0 when the bug
> >>> triggers (and i_head_snapc != NULL).
> >>>
> >>
> >> Thanks, which test triggers this bug?
> >
> > That's generic/013. It usually triggers after a few hours of running it
> > in a loop (I'm using a vstart cluster for that).
> >
> >>
> >> I searched that code, found we may fail to cleanup i_head_snap in two
> >> places. One is in ceph_queue_cap_snap, Another is in
> >> remove_session_caps_cb().
> >
> > Ah, great! I spent a lot of time looking but I couldn't really find it.
> > My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
> > and that the corresponding ceph_put_snap_context would be done in
> > handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
> > MDS.
>
> Something that I didn't said explicitly is that I *know* that the
> unbalanced ceph_get_snap_context() is the one in function
> ceph_queue_cap_snap(). I know this for sure because I've managed to
> reproduce the issue several times with an instrumented kernel (with
> tracepoints) that allowed me to keep track of all ceph_snap_context
> operations (create, get, put, delete).
>

Sorry for the delay. please try

diff --git a/fs/ceph/mds_client.c b/fs/ceph/mds_client.c
index 45f061e47721..5bc30e0d252a 100644
--- a/fs/ceph/mds_client.c
+++ b/fs/ceph/mds_client.c
@@ -1420,6 +1420,15 @@ static int remove_session_caps_cb(struct inode
*inode, struct ceph_cap *cap,
list_add(&ci->i_prealloc_cap_flush->i_list, &to_remove);
ci->i_prealloc_cap_flush = NULL;
}
+
+ if (drop &&
+ ci->i_wrbuffer_ref_head == 0 &&
+ ci->i_wr_ref == 0 &&
+ ci->i_dirty_caps == 0 &&
+ ci->i_flushing_caps == 0) {
+ ceph_put_snap_context(ci->i_head_snapc);
+ ci->i_head_snapc = NULL;
+ }
}
spin_unlock(&ci->i_ceph_lock);
while (!list_empty(&to_remove)) {
diff --git a/fs/ceph/snap.c b/fs/ceph/snap.c
index 89aa37fa0f84..0af4b3d32122 100644
--- a/fs/ceph/snap.c
+++ b/fs/ceph/snap.c
@@ -572,7 +572,12 @@ void ceph_queue_cap_snap(struct ceph_inode_info *ci)
old_snapc = NULL;

update_snapc:
- if (ci->i_head_snapc) {
+ if (ci->i_wrbuffer_ref_head == 0 &&
+ ci->i_wr_ref == 0 &&
+ ci->i_dirty_caps == 0 &&
+ ci->i_flushing_caps == 0) {
+ ci->i_head_snapc = NULL;
+ } else {
ci->i_head_snapc = ceph_get_snap_context(new_snapc);
dout(" new snapc is %p\n", new_snapc);
}



> Cheers,
> --
> Luis

2019-04-03 09:50:27

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

"Yan, Zheng" <[email protected]> writes:

> On Fri, Mar 22, 2019 at 6:04 PM Luis Henriques <[email protected]> wrote:
>>
>> Luis Henriques <[email protected]> writes:
>>
>> > "Yan, Zheng" <[email protected]> writes:
>> >
>> >> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
>> >>>
>> >>> "Yan, Zheng" <[email protected]> writes:
>> >>>
>> >>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>> >>> >>
>> >>> >> "Yan, Zheng" <[email protected]> writes:
>> >>> >>
>> >>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>> >>> >> >>
>> >>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>> >>> >> >>
>> >>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
>> >>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>> >>> >> >> hex dump (first 32 bytes):
>> >>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>> >>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>> >>> >> >> backtrace:
>> >>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>> >>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>> >>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>> >>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>> >>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>> >>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
>> >>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>> >>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>> >>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>> >>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
>> >>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>> >>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
>> >>> >> >>
>> >>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
>> >>> >> >> for example, umounting the filesystem. In that case, we can simply drop
>> >>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>> >>> >> >>
>> >>> >> >> Link: https://tracker.ceph.com/issues/38224
>> >>> >> >> Cc: [email protected]
>> >>> >> >> Signed-off-by: Luis Henriques <[email protected]>
>> >>> >> >> ---
>> >>> >> >> fs/ceph/caps.c | 7 +++++++
>> >>> >> >> 1 file changed, 7 insertions(+)
>> >>> >> >>
>> >>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>> >>> >> >> index 36a8dc699448..208f4dc6f574 100644
>> >>> >> >> --- a/fs/ceph/caps.c
>> >>> >> >> +++ b/fs/ceph/caps.c
>> >>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>> >>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >>> >> >> {
>> >>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
>> >>> >> >> +
>> >>> >> >> spin_lock(&realm->inodes_with_caps_lock);
>> >>> >> >> list_del_init(&ci->i_snap_realm_item);
>> >>> >> >> ci->i_snap_realm_counter++;
>> >>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>> >>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
>> >>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>> >>> >> >> realm);
>> >>> >> >> + /*
>> >>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>> >>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
>> >>> >> >> + * ceph_snap_context and we need to drop it.
>> >>> >> >> + */
>> >>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
>> >>> >> >> }
>> >>> >> >>
>> >>> >> >> /*
>> >>> >> >
>> >>> >> > This does not seem right. i_head_snapc is cleared when
>> >>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
>> >>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
>> >>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
>> >>> >> > should be in iterate_session_caps()
>> >>> >> >
>> >>> >>
>> >>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
>> >>> >> seems to execute fine, except from the memory leak.
>> >>> >>
>> >>> >> It's very difficult to reproduce this issue, but last time I managed to
>> >>> >> get this memory leak to trigger I actually had some debugging code in
>> >>> >> drop_inode_snap_realm, something like:
>> >>> >>
>> >>> >> if (ci->i_head_snapc)
>> >>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
>> >>> >
>> >>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
>> >>> > i_flushing_caps. and try reproducing it again.
>> >>> >
>> >>>
>> >>> Ok, it took me a few hours, but I managed to reproduce the bug, with
>> >>> those extra printks. All those values are set to 0 when the bug
>> >>> triggers (and i_head_snapc != NULL).
>> >>>
>> >>
>> >> Thanks, which test triggers this bug?
>> >
>> > That's generic/013. It usually triggers after a few hours of running it
>> > in a loop (I'm using a vstart cluster for that).
>> >
>> >>
>> >> I searched that code, found we may fail to cleanup i_head_snap in two
>> >> places. One is in ceph_queue_cap_snap, Another is in
>> >> remove_session_caps_cb().
>> >
>> > Ah, great! I spent a lot of time looking but I couldn't really find it.
>> > My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
>> > and that the corresponding ceph_put_snap_context would be done in
>> > handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
>> > MDS.
>>
>> Something that I didn't said explicitly is that I *know* that the
>> unbalanced ceph_get_snap_context() is the one in function
>> ceph_queue_cap_snap(). I know this for sure because I've managed to
>> reproduce the issue several times with an instrumented kernel (with
>> tracepoints) that allowed me to keep track of all ceph_snap_context
>> operations (create, get, put, delete).
>>
>
> Sorry for the delay. please try

Thank you, I'll try the test with your patch applied and see if I still
see the issue (obviously, this will take some time as it takes several
hours for the bug to trigger).

Cheers,
--
Luis


>
> diff --git a/fs/ceph/mds_client.c b/fs/ceph/mds_client.c
> index 45f061e47721..5bc30e0d252a 100644
> --- a/fs/ceph/mds_client.c
> +++ b/fs/ceph/mds_client.c
> @@ -1420,6 +1420,15 @@ static int remove_session_caps_cb(struct inode
> *inode, struct ceph_cap *cap,
> list_add(&ci->i_prealloc_cap_flush->i_list, &to_remove);
> ci->i_prealloc_cap_flush = NULL;
> }
> +
> + if (drop &&
> + ci->i_wrbuffer_ref_head == 0 &&
> + ci->i_wr_ref == 0 &&
> + ci->i_dirty_caps == 0 &&
> + ci->i_flushing_caps == 0) {
> + ceph_put_snap_context(ci->i_head_snapc);
> + ci->i_head_snapc = NULL;
> + }
> }
> spin_unlock(&ci->i_ceph_lock);
> while (!list_empty(&to_remove)) {
> diff --git a/fs/ceph/snap.c b/fs/ceph/snap.c
> index 89aa37fa0f84..0af4b3d32122 100644
> --- a/fs/ceph/snap.c
> +++ b/fs/ceph/snap.c
> @@ -572,7 +572,12 @@ void ceph_queue_cap_snap(struct ceph_inode_info *ci)
> old_snapc = NULL;
>
> update_snapc:
> - if (ci->i_head_snapc) {
> + if (ci->i_wrbuffer_ref_head == 0 &&
> + ci->i_wr_ref == 0 &&
> + ci->i_dirty_caps == 0 &&
> + ci->i_flushing_caps == 0) {
> + ci->i_head_snapc = NULL;
> + } else {
> ci->i_head_snapc = ceph_get_snap_context(new_snapc);
> dout(" new snapc is %p\n", new_snapc);
> }
>
>
>
>> Cheers,
>> --
>> Luis
>

2019-04-16 13:31:07

by Luis Henriques

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

Luis Henriques <[email protected]> writes:

> "Yan, Zheng" <[email protected]> writes:
>
>> On Fri, Mar 22, 2019 at 6:04 PM Luis Henriques <[email protected]> wrote:
>>>
>>> Luis Henriques <[email protected]> writes:
>>>
>>> > "Yan, Zheng" <[email protected]> writes:
>>> >
>>> >> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
>>> >>>
>>> >>> "Yan, Zheng" <[email protected]> writes:
>>> >>>
>>> >>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
>>> >>> >>
>>> >>> >> "Yan, Zheng" <[email protected]> writes:
>>> >>> >>
>>> >>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
>>> >>> >> >>
>>> >>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
>>> >>> >> >>
>>> >>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
>>> >>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
>>> >>> >> >> hex dump (first 32 bytes):
>>> >>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
>>> >>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
>>> >>> >> >> backtrace:
>>> >>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
>>> >>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
>>> >>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
>>> >>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
>>> >>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
>>> >>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
>>> >>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
>>> >>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
>>> >>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
>>> >>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
>>> >>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
>>> >>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
>>> >>> >> >>
>>> >>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
>>> >>> >> >> for example, umounting the filesystem. In that case, we can simply drop
>>> >>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
>>> >>> >> >>
>>> >>> >> >> Link: https://tracker.ceph.com/issues/38224
>>> >>> >> >> Cc: [email protected]
>>> >>> >> >> Signed-off-by: Luis Henriques <[email protected]>
>>> >>> >> >> ---
>>> >>> >> >> fs/ceph/caps.c | 7 +++++++
>>> >>> >> >> 1 file changed, 7 insertions(+)
>>> >>> >> >>
>>> >>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
>>> >>> >> >> index 36a8dc699448..208f4dc6f574 100644
>>> >>> >> >> --- a/fs/ceph/caps.c
>>> >>> >> >> +++ b/fs/ceph/caps.c
>>> >>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
>>> >>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>>> >>> >> >> {
>>> >>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
>>> >>> >> >> +
>>> >>> >> >> spin_lock(&realm->inodes_with_caps_lock);
>>> >>> >> >> list_del_init(&ci->i_snap_realm_item);
>>> >>> >> >> ci->i_snap_realm_counter++;
>>> >>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
>>> >>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
>>> >>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
>>> >>> >> >> realm);
>>> >>> >> >> + /*
>>> >>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
>>> >>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
>>> >>> >> >> + * ceph_snap_context and we need to drop it.
>>> >>> >> >> + */
>>> >>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
>>> >>> >> >> }
>>> >>> >> >>
>>> >>> >> >> /*
>>> >>> >> >
>>> >>> >> > This does not seem right. i_head_snapc is cleared when
>>> >>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
>>> >>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
>>> >>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
>>> >>> >> > should be in iterate_session_caps()
>>> >>> >> >
>>> >>> >>
>>> >>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
>>> >>> >> seems to execute fine, except from the memory leak.
>>> >>> >>
>>> >>> >> It's very difficult to reproduce this issue, but last time I managed to
>>> >>> >> get this memory leak to trigger I actually had some debugging code in
>>> >>> >> drop_inode_snap_realm, something like:
>>> >>> >>
>>> >>> >> if (ci->i_head_snapc)
>>> >>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
>>> >>> >
>>> >>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
>>> >>> > i_flushing_caps. and try reproducing it again.
>>> >>> >
>>> >>>
>>> >>> Ok, it took me a few hours, but I managed to reproduce the bug, with
>>> >>> those extra printks. All those values are set to 0 when the bug
>>> >>> triggers (and i_head_snapc != NULL).
>>> >>>
>>> >>
>>> >> Thanks, which test triggers this bug?
>>> >
>>> > That's generic/013. It usually triggers after a few hours of running it
>>> > in a loop (I'm using a vstart cluster for that).
>>> >
>>> >>
>>> >> I searched that code, found we may fail to cleanup i_head_snap in two
>>> >> places. One is in ceph_queue_cap_snap, Another is in
>>> >> remove_session_caps_cb().
>>> >
>>> > Ah, great! I spent a lot of time looking but I couldn't really find it.
>>> > My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
>>> > and that the corresponding ceph_put_snap_context would be done in
>>> > handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
>>> > MDS.
>>>
>>> Something that I didn't said explicitly is that I *know* that the
>>> unbalanced ceph_get_snap_context() is the one in function
>>> ceph_queue_cap_snap(). I know this for sure because I've managed to
>>> reproduce the issue several times with an instrumented kernel (with
>>> tracepoints) that allowed me to keep track of all ceph_snap_context
>>> operations (create, get, put, delete).
>>>
>>
>> Sorry for the delay. please try
>
> Thank you, I'll try the test with your patch applied and see if I still
> see the issue (obviously, this will take some time as it takes several
> hours for the bug to trigger).

Ok, after several hours of running it, I confirm that I was not able to
reproduce the issue with your patch. It took me a while because my test
environment changed a bit and I was failing to reproduce it even without
the patch.

Cheers,
--
Luis

2019-04-18 03:39:10

by Yan, Zheng

[permalink] [raw]
Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc

On Tue, Apr 16, 2019 at 9:30 PM Luis Henriques <[email protected]> wrote:
>
> Luis Henriques <[email protected]> writes:
>
> > "Yan, Zheng" <[email protected]> writes:
> >
> >> On Fri, Mar 22, 2019 at 6:04 PM Luis Henriques <[email protected]> wrote:
> >>>
> >>> Luis Henriques <[email protected]> writes:
> >>>
> >>> > "Yan, Zheng" <[email protected]> writes:
> >>> >
> >>> >> On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques <[email protected]> wrote:
> >>> >>>
> >>> >>> "Yan, Zheng" <[email protected]> writes:
> >>> >>>
> >>> >>> > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques <[email protected]> wrote:
> >>> >>> >>
> >>> >>> >> "Yan, Zheng" <[email protected]> writes:
> >>> >>> >>
> >>> >>> >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques <[email protected]> wrote:
> >>> >>> >> >>
> >>> >>> >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013:
> >>> >>> >> >>
> >>> >>> >> >> unreferenced object 0xffff8881fccca940 (size 32):
> >>> >>> >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s)
> >>> >>> >> >> hex dump (first 32 bytes):
> >>> >>> >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................
> >>> >>> >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................
> >>> >>> >> >> backtrace:
> >>> >>> >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0
> >>> >>> >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90
> >>> >>> >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90
> >>> >>> >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610
> >>> >>> >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3
> >>> >>> >> >> [<00000000fc287b83>] dispatch+0x362/0x176c
> >>> >>> >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0
> >>> >>> >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400
> >>> >>> >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0
> >>> >>> >> >> [<00000000b593e4b3>] kthread+0x112/0x130
> >>> >>> >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40
> >>> >>> >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff
> >>> >>> >> >>
> >>> >>> >> >> It looks like it is possible that we miss a flush_ack from the MDS when,
> >>> >>> >> >> for example, umounting the filesystem. In that case, we can simply drop
> >>> >>> >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap().
> >>> >>> >> >>
> >>> >>> >> >> Link: https://tracker.ceph.com/issues/38224
> >>> >>> >> >> Cc: [email protected]
> >>> >>> >> >> Signed-off-by: Luis Henriques <[email protected]>
> >>> >>> >> >> ---
> >>> >>> >> >> fs/ceph/caps.c | 7 +++++++
> >>> >>> >> >> 1 file changed, 7 insertions(+)
> >>> >>> >> >>
> >>> >>> >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c
> >>> >>> >> >> index 36a8dc699448..208f4dc6f574 100644
> >>> >>> >> >> --- a/fs/ceph/caps.c
> >>> >>> >> >> +++ b/fs/ceph/caps.c
> >>> >>> >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode)
> >>> >>> >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >>> >>> >> >> {
> >>> >>> >> >> struct ceph_snap_realm *realm = ci->i_snap_realm;
> >>> >>> >> >> +
> >>> >>> >> >> spin_lock(&realm->inodes_with_caps_lock);
> >>> >>> >> >> list_del_init(&ci->i_snap_realm_item);
> >>> >>> >> >> ci->i_snap_realm_counter++;
> >>> >>> >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci)
> >>> >>> >> >> spin_unlock(&realm->inodes_with_caps_lock);
> >>> >>> >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc,
> >>> >>> >> >> realm);
> >>> >>> >> >> + /*
> >>> >>> >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a
> >>> >>> >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the
> >>> >>> >> >> + * ceph_snap_context and we need to drop it.
> >>> >>> >> >> + */
> >>> >>> >> >> + ceph_put_snap_context(ci->i_head_snapc);
> >>> >>> >> >> }
> >>> >>> >> >>
> >>> >>> >> >> /*
> >>> >>> >> >
> >>> >>> >> > This does not seem right. i_head_snapc is cleared when
> >>> >>> >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 &&
> >>> >>> >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm.
> >>> >>> >> > Did you see 'reconnect denied' during the test? If you did, the fix
> >>> >>> >> > should be in iterate_session_caps()
> >>> >>> >> >
> >>> >>> >>
> >>> >>> >> No, I didn't saw any 'reconnect denied' in the test. The test actually
> >>> >>> >> seems to execute fine, except from the memory leak.
> >>> >>> >>
> >>> >>> >> It's very difficult to reproduce this issue, but last time I managed to
> >>> >>> >> get this memory leak to trigger I actually had some debugging code in
> >>> >>> >> drop_inode_snap_realm, something like:
> >>> >>> >>
> >>> >>> >> if (ci->i_head_snapc)
> >>> >>> >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc);
> >>> >>> >
> >>> >>> > please add code that prints i_wrbuffer_ref_head, i_dirty_caps,
> >>> >>> > i_flushing_caps. and try reproducing it again.
> >>> >>> >
> >>> >>>
> >>> >>> Ok, it took me a few hours, but I managed to reproduce the bug, with
> >>> >>> those extra printks. All those values are set to 0 when the bug
> >>> >>> triggers (and i_head_snapc != NULL).
> >>> >>>
> >>> >>
> >>> >> Thanks, which test triggers this bug?
> >>> >
> >>> > That's generic/013. It usually triggers after a few hours of running it
> >>> > in a loop (I'm using a vstart cluster for that).
> >>> >
> >>> >>
> >>> >> I searched that code, found we may fail to cleanup i_head_snap in two
> >>> >> places. One is in ceph_queue_cap_snap, Another is in
> >>> >> remove_session_caps_cb().
> >>> >
> >>> > Ah, great! I spent a lot of time looking but I couldn't really find it.
> >>> > My bet was that ceph_queue_cap_snap was doing the ceph_get_snap_context
> >>> > and that the corresponding ceph_put_snap_context would be done in
> >>> > handle_cap_flush_ack. That's why I mentioned the missing flush_ack from
> >>> > MDS.
> >>>
> >>> Something that I didn't said explicitly is that I *know* that the
> >>> unbalanced ceph_get_snap_context() is the one in function
> >>> ceph_queue_cap_snap(). I know this for sure because I've managed to
> >>> reproduce the issue several times with an instrumented kernel (with
> >>> tracepoints) that allowed me to keep track of all ceph_snap_context
> >>> operations (create, get, put, delete).
> >>>
> >>
> >> Sorry for the delay. please try
> >
> > Thank you, I'll try the test with your patch applied and see if I still
> > see the issue (obviously, this will take some time as it takes several
> > hours for the bug to trigger).
>
> Ok, after several hours of running it, I confirm that I was not able to
> reproduce the issue with your patch. It took me a while because my test
> environment changed a bit and I was failing to reproduce it even without
> the patch.
>

Thanks, I pushed the patch to testing branch.

Yan, Zheng

> Cheers,
> --
> Luis