Received: by 2002:ac0:bc90:0:0:0:0:0 with SMTP id a16csp986567img; Mon, 18 Mar 2019 20:17:44 -0700 (PDT) X-Google-Smtp-Source: APXvYqx9B7RyCqsjIVc3qCEneSy+pLe/mCFvRzK4c+nxiuOKxOGA5KAXrvWW5fjjFbx7T/86l7Ih X-Received: by 2002:a63:f850:: with SMTP id v16mr3983212pgj.448.1552965464907; Mon, 18 Mar 2019 20:17:44 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1552965464; cv=none; d=google.com; s=arc-20160816; b=CUjXiPUn7onD/NU4YVM623wvx/Gjtfss92ZZTc7NpTeWyKOvwyvG3MUTclKZC0Gw2u L4Y2Bp5CHiBcWgl3ClGMMc2bk7XZP2HzbLABryynmoSdEy5Ay8/dkMRzUTgG7KT5RWh9 ONDxDLUT0m80yC5AtH/jCNWiLa/3ysr9GD00Ku2Whch72M7b8lje/3OoqjnLDMapVCtD 1cj6zQYTKyrCaSWY9lUwtQV+CY7GZSbOM4afwbrUmCGRdkfNhV1rXBXcT4N8JdmXlm2k 8/NwXNMJBZMk+OUl7cXFeUtvmkz07d+zDtkxCR7fQAWrsUsb/U4GIOoCxiBAJRcyPp8q ecRA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=jGGSnafaNBpdSiNLLCqzapKuD/ZrthMIzJSfiQIdF54=; b=BtCZ0+N0LjqM0H3oHmhLbO8ORksi/UWJuVcA3YsjUqTr1g9SfpUhw6UZpU9YF8JRWc uDP9PP+aubLpmsackGiNJ8v1dDaMR+vahNYbb4DfNwBVBRcEgQq+ws/vI1SAMvnF9iEU LIpeYETcvKkVjyfBg70LWvYBlllm9OuS3MjP0Ve6rbQ4bOCk0jc2IzY3H63yV+Z2jN5h jEsxzUxVGd8SX+Cht7p6pjj+tNKUoDqGI8f5RAm9Li4+4qanBY4zkCfxH7CdWij95DYG +TehAC1PDHZFOvSV/nkDR4a1syVdZxY8reUQdOQ6IxsWEUVJPNIWmgmeI0BGY4NaXoF1 yeGw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=G1buR+fh; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id g4si10448046plt.215.2019.03.18.20.17.29; Mon, 18 Mar 2019 20:17:44 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=G1buR+fh; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727532AbfCSDNs (ORCPT + 99 others); Mon, 18 Mar 2019 23:13:48 -0400 Received: from mail-qt1-f193.google.com ([209.85.160.193]:36673 "EHLO mail-qt1-f193.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726788AbfCSDNs (ORCPT ); Mon, 18 Mar 2019 23:13:48 -0400 Received: by mail-qt1-f193.google.com with SMTP id y36so10769793qtb.3; Mon, 18 Mar 2019 20:13:47 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=jGGSnafaNBpdSiNLLCqzapKuD/ZrthMIzJSfiQIdF54=; b=G1buR+fh/psy/75kjeCHD+SHsR7M0Pr0aXbMeko6MfLOZNmUQCzPkkj3QCJ7UM4Bwt EmvnH3CUwl+zzk/u2cz0i384UTd7sDQTphPs0s1cr+rUGbGO4C4daiTRHjQQe7cuJlFm dPFi6NkD4PW5qEbTmwNo/zyGLPd8jz2RgJztukXHlzhFAa06qlb7di7xUqErcGx19BDL STKnMmIVof7SLcRPGbPCIfRW94iN7Lgw3I1sbFryoERiGUARVSaCpUawvswUsiWfyxPX IqWT48guKvRX9TrYfcPYUCfk7yX/II72/KSQ6u84nCSOoYrVJiO9l+CWxMsmnRrx6ugw BlqQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=jGGSnafaNBpdSiNLLCqzapKuD/ZrthMIzJSfiQIdF54=; b=DOtNWKQNS8NR9cZrFm9xmQ+jMmHv8jEyri2f3wYs51mhf9zoYzdNRVY6j2mg+7QGtD U8N3t0bE4CrIvlC5JHkoHY5/Uehj/NA/OCaoax2FKUwHTYqYjs/UCUpV9Qvj4eol5Nlh AFiONB7AqTAnieZEpxsDvMYvQQPd0hauA1NLnS8rdFRHis2r9oUvgMnZPQAjx2dsVrbL 0Ui53ywIAhW1ClhYqFTxGH8E0uVrpW0M3Q0EmdUknG9xHJdJaxRQCFSLmBu+wLZoJBZN a9ddP03pBAp74PooMEFppfEmMFzhugbUdB1w8L1xT1ewZUgBoA2wOV9rPS3GUcRWsbdD 7NfQ== X-Gm-Message-State: APjAAAWyGT9ZffaJ2lA+DcjDfugH+KgUSc0h/hcA2qcAq+TGKMERY/rb y34VcK0+AMMKwiKlsiGIQC7z5HDHacFVBkemWwI= X-Received: by 2002:a0c:c785:: with SMTP id k5mr469951qvj.0.1552965227021; Mon, 18 Mar 2019 20:13:47 -0700 (PDT) MIME-Version: 1.0 References: <20190315111107.15154-1-lhenriques@suse.com> <87r2b4zd2q.fsf@suse.com> <87lg1c17b2.fsf@suse.com> In-Reply-To: <87lg1c17b2.fsf@suse.com> From: "Yan, Zheng" Date: Tue, 19 Mar 2019 11:13:35 +0800 Message-ID: Subject: Re: [PATCH] ceph: Fix a memory leak in ci->i_head_snapc To: Luis Henriques Cc: "Yan, Zheng" , Sage Weil , Ilya Dryomov , ceph-devel , Linux Kernel Mailing List , stable@vger.kernel.org Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Mar 19, 2019 at 12:22 AM Luis Henriques wrote: > > "Yan, Zheng" writes: > > > On Mon, Mar 18, 2019 at 6:33 PM Luis Henriques wrote: > >> > >> "Yan, Zheng" writes: > >> > >> > On Fri, Mar 15, 2019 at 7:13 PM Luis Henriques wrote: > >> >> > >> >> I'm occasionally seeing a kmemleak warning in xfstest generic/013: > >> >> > >> >> unreferenced object 0xffff8881fccca940 (size 32): > >> >> comm "kworker/0:1", pid 12, jiffies 4295005883 (age 130.648s) > >> >> hex dump (first 32 bytes): > >> >> 01 00 00 00 00 00 00 00 01 00 00 00 00 00 00 00 ................ > >> >> 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ > >> >> backtrace: > >> >> [<00000000d741a1ea>] build_snap_context+0x5b/0x2a0 > >> >> [<0000000021a00533>] rebuild_snap_realms+0x27/0x90 > >> >> [<00000000ac538600>] rebuild_snap_realms+0x42/0x90 > >> >> [<000000000e955fac>] ceph_update_snap_trace+0x2ee/0x610 > >> >> [<00000000a9550416>] ceph_handle_snap+0x317/0x5f3 > >> >> [<00000000fc287b83>] dispatch+0x362/0x176c > >> >> [<00000000a312c741>] ceph_con_workfn+0x9ce/0x2cf0 > >> >> [<000000004168e3a9>] process_one_work+0x1d4/0x400 > >> >> [<000000002188e9e7>] worker_thread+0x2d/0x3c0 > >> >> [<00000000b593e4b3>] kthread+0x112/0x130 > >> >> [<00000000a8587dca>] ret_from_fork+0x35/0x40 > >> >> [<00000000ba1c9c1d>] 0xffffffffffffffff > >> >> > >> >> It looks like it is possible that we miss a flush_ack from the MDS when, > >> >> for example, umounting the filesystem. In that case, we can simply drop > >> >> the reference to the ceph_snap_context obtained in ceph_queue_cap_snap(). > >> >> > >> >> Link: https://tracker.ceph.com/issues/38224 > >> >> Cc: stable@vger.kernel.org > >> >> Signed-off-by: Luis Henriques > >> >> --- > >> >> fs/ceph/caps.c | 7 +++++++ > >> >> 1 file changed, 7 insertions(+) > >> >> > >> >> diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c > >> >> index 36a8dc699448..208f4dc6f574 100644 > >> >> --- a/fs/ceph/caps.c > >> >> +++ b/fs/ceph/caps.c > >> >> @@ -1054,6 +1054,7 @@ int ceph_is_any_caps(struct inode *inode) > >> >> static void drop_inode_snap_realm(struct ceph_inode_info *ci) > >> >> { > >> >> struct ceph_snap_realm *realm = ci->i_snap_realm; > >> >> + > >> >> spin_lock(&realm->inodes_with_caps_lock); > >> >> list_del_init(&ci->i_snap_realm_item); > >> >> ci->i_snap_realm_counter++; > >> >> @@ -1063,6 +1064,12 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci) > >> >> spin_unlock(&realm->inodes_with_caps_lock); > >> >> ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc, > >> >> realm); > >> >> + /* > >> >> + * ci->i_head_snapc should be NULL, but we may still be waiting for a > >> >> + * flush_ack from the MDS. In that case, we still hold a ref for the > >> >> + * ceph_snap_context and we need to drop it. > >> >> + */ > >> >> + ceph_put_snap_context(ci->i_head_snapc); > >> >> } > >> >> > >> >> /* > >> > > >> > This does not seem right. i_head_snapc is cleared when > >> > (ci->i_wrbuffer_ref_head == 0 && ci->i_dirty_caps == 0 && > >> > ci->i_flushing_caps == 0) . Nothing do with dropping ci->i_snap_realm. > >> > Did you see 'reconnect denied' during the test? If you did, the fix > >> > should be in iterate_session_caps() > >> > > >> > >> No, I didn't saw any 'reconnect denied' in the test. The test actually > >> seems to execute fine, except from the memory leak. > >> > >> It's very difficult to reproduce this issue, but last time I managed to > >> get this memory leak to trigger I actually had some debugging code in > >> drop_inode_snap_realm, something like: > >> > >> if (ci->i_head_snapc) > >> printk("i_head_snapc: 0x%px\n", ci->i_head_snapc); > > > > please add code that prints i_wrbuffer_ref_head, i_dirty_caps, > > i_flushing_caps. and try reproducing it again. > > > > Ok, it took me a few hours, but I managed to reproduce the bug, with > those extra printks. All those values are set to 0 when the bug > triggers (and i_head_snapc != NULL). > Thanks, which test triggers this bug? I searched that code, found we may fail to cleanup i_head_snap in two places. One is in ceph_queue_cap_snap, Another is in remove_session_caps_cb(). > Cheers, > -- > Luis > > > > > >> > >> This printk was only executed when the bug triggered (during a > >> filesystem umount) and the address shown was the same as in the kmemleak > >> warning. > >> > >> After spending some time looking, I assumed this to be a missing call to > >> handle_cap_flush_ack, which would do the i_head_snapc cleanup. > >> > >> Cheers, > >> -- > >> Luis > >