Received: by 2002:a05:6602:2086:0:0:0:0 with SMTP id a6csp4635994ioa; Wed, 27 Apr 2022 07:59:51 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxAvx8sIyw+w9JcJkGKNBVn55nqE4MK1xN/+zyanfz96MeOrP9y7ARcO5ZXreRVGOhnLhtM X-Received: by 2002:a17:90a:62c7:b0:1da:2c51:943e with SMTP id k7-20020a17090a62c700b001da2c51943emr5306639pjs.208.1651071591679; Wed, 27 Apr 2022 07:59:51 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1651071591; cv=none; d=google.com; s=arc-20160816; b=ArtlQHOBeZx02D8yuOy7o1K11lyHkpVIK0Qhn+UugqtKzQIqvFBisEurJbTEQjJl7a XQRL5x+pPcDaIVmaFooDadZpFFEfB9hJpa/VRjfRhbUtc4H26L81SZGJx6VYUHVNn/51 P0Dl+1mhs5Bbx2xSGcv6gQDfmjm/8c4kq08b0UW3kX0tuhq95rAdnoImhTvswTGifESg Nm/1AsSQGjPH2rM20DuAqgyPhm3pAvrMFjbqkdFOqDg0rEsYv+BMZWoC2NEMbimcykU2 BtKgWc/ySs1JDaL556PLVwY+1mITGiwMkyD9trcZ1m9RE1OfppxehZ99MdcqLBRCMIT1 er5Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :message-id:date:subject:cc:to:from:dkim-signature:dkim-signature; bh=i/KJP+44iudIC+WcuC2nIb/z3EQuXQtWQd5UTDcPQTs=; b=TJMg5E3OW/fofL6cjHuzzircBHntykwpjHMm27fRNvH47VjJgTRUoZLH2GUvSLmNH7 I6nCH0C4iAtEsRsL8DZPtF9rV6q8qHxgQhPbO5DYGCv6ibeBBeg0BxcrSz+bdyoGL4Hw y0muMjzgGyf8yC6xMtzjeVdWsourbr7MJOkql3KxG1Su0vl61pzsSMUUfGGNLiH9z6bT 8soafLFhzQ8I30/Roc8oPknmdABYjbpBHU/VJdGlhGDuBv5KEBv5i5OVHT3uCWNpxrI4 Fo8HGfal4i8f8L2MNshTzcJ65QoP985ffrrz9Jj/kT4YHr5e4SkMPVGOJJDhuF9mFiqf n2mw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@suse.de header.s=susede2_rsa header.b=RyET7+oN; dkim=neutral (no key) header.i=@suse.de header.s=susede2_ed25519 header.b="jfY/6Nt5"; spf=softfail (google.com: domain of transitioning linux-kernel-owner@vger.kernel.org does not designate 23.128.96.19 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=suse.de Return-Path: Received: from lindbergh.monkeyblade.net (lindbergh.monkeyblade.net. [23.128.96.19]) by mx.google.com with ESMTPS id e191-20020a6369c8000000b003c1425344cdsi1475812pgc.584.2022.04.27.07.59.51 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 27 Apr 2022 07:59:51 -0700 (PDT) Received-SPF: softfail (google.com: domain of transitioning linux-kernel-owner@vger.kernel.org does not designate 23.128.96.19 as permitted sender) client-ip=23.128.96.19; Authentication-Results: mx.google.com; dkim=pass header.i=@suse.de header.s=susede2_rsa header.b=RyET7+oN; dkim=neutral (no key) header.i=@suse.de header.s=susede2_ed25519 header.b="jfY/6Nt5"; spf=softfail (google.com: domain of transitioning linux-kernel-owner@vger.kernel.org does not designate 23.128.96.19 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=suse.de Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 75DA3263A; Wed, 27 Apr 2022 07:32:49 -0700 (PDT) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238027AbiD0Ofw (ORCPT + 99 others); Wed, 27 Apr 2022 10:35:52 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37412 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S238010AbiD0Ofs (ORCPT ); Wed, 27 Apr 2022 10:35:48 -0400 Received: from smtp-out1.suse.de (smtp-out1.suse.de [195.135.220.28]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E827C1B9; Wed, 27 Apr 2022 07:32:36 -0700 (PDT) Received: from imap2.suse-dmz.suse.de (imap2.suse-dmz.suse.de [192.168.254.74]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature ECDSA (P-521) server-digest SHA512) (No client certificate requested) by smtp-out1.suse.de (Postfix) with ESMTPS id 829BF21122; Wed, 27 Apr 2022 14:32:35 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=suse.de; s=susede2_rsa; t=1651069955; h=from:from:reply-to:date:date:message-id:message-id:to:to:cc:cc: mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding; bh=i/KJP+44iudIC+WcuC2nIb/z3EQuXQtWQd5UTDcPQTs=; b=RyET7+oNDgi2ZXLUior/Ne6UXO3HdPEr07aDMFH4PtpAOpP33zyVDEVHUp3HW8avYVOozI rN7+AWuDvBTed8b2tAe3vggF78CQ3Ie9N84F/fEa4pTI58mhVh78XlNp2EUcIWo64jHukY xkBUnJMQxfVKEDTVQSNIExJCxjrvkXI= DKIM-Signature: v=1; a=ed25519-sha256; c=relaxed/relaxed; d=suse.de; s=susede2_ed25519; t=1651069955; h=from:from:reply-to:date:date:message-id:message-id:to:to:cc:cc: mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding; bh=i/KJP+44iudIC+WcuC2nIb/z3EQuXQtWQd5UTDcPQTs=; b=jfY/6Nt5pZVzjT3TDrhJQmX8pEeTBvelR36nJbujlnPKjT1eyNCefX46jFb5D/rK4IE0kB Z85PtwNW9yJaAFCg== Received: from imap2.suse-dmz.suse.de (imap2.suse-dmz.suse.de [192.168.254.74]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature ECDSA (P-521) server-digest SHA512) (No client certificate requested) by imap2.suse-dmz.suse.de (Postfix) with ESMTPS id 1525013A39; Wed, 27 Apr 2022 14:32:35 +0000 (UTC) Received: from dovecot-director2.suse.de ([192.168.254.65]) by imap2.suse-dmz.suse.de with ESMTPSA id xUJlAgNUaWIKeQAAMHmgww (envelope-from ); Wed, 27 Apr 2022 14:32:35 +0000 Received: from localhost (brahms.olymp [local]) by brahms.olymp (OpenSMTPD) with ESMTPA id 8dfb1ff8; Wed, 27 Apr 2022 14:33:05 +0000 (UTC) From: =?UTF-8?q?Lu=C3=ADs=20Henriques?= To: Jeff Layton , Xiubo Li , Ilya Dryomov Cc: ceph-devel@vger.kernel.org, linux-kernel@vger.kernel.org, =?UTF-8?q?Lu=C3=ADs=20Henriques?= , Ryan Taylor Subject: [PATCH v2] ceph: fix statfs for subdir mounts Date: Wed, 27 Apr 2022 15:33:03 +0100 Message-Id: <20220427143303.950-1-lhenriques@suse.de> MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-2.0 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,RDNS_NONE,SPF_HELO_NONE autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org When doing a mount using as base a directory that has 'max_bytes' quotas statfs uses that value as the total; if a subdirectory is used instead, the same 'max_bytes' too in statfs, unless there is another quota set. Unfortunately, if this subdirectory only has the 'max_files' quota set, then statfs uses the filesystem total. Fix this by making sure we only lookup realms that contain the 'max_bytes' quota. Link: https://tracker.ceph.com/issues/55090 Cc: Ryan Taylor Signed-off-by: Luís Henriques --- As I mentioned in v1, I do *not* think this really fixes the tracker above, as the bug reporter never mentioned setting quotas in the subdir. Changes since v1: Moved some more logic into __ceph_has_any_quota() function. fs/ceph/inode.c | 2 +- fs/ceph/quota.c | 19 +++++++++++-------- fs/ceph/super.h | 28 ++++++++++++++++++++++++---- 3 files changed, 36 insertions(+), 13 deletions(-) diff --git a/fs/ceph/inode.c b/fs/ceph/inode.c index 5de7bb9048b7..4b7406d6fbe4 100644 --- a/fs/ceph/inode.c +++ b/fs/ceph/inode.c @@ -691,7 +691,7 @@ void ceph_evict_inode(struct inode *inode) __ceph_remove_caps(ci); - if (__ceph_has_any_quota(ci)) + if (__ceph_has_any_quota(ci, QUOTA_GET_ANY)) ceph_adjust_quota_realms_count(inode, false); /* diff --git a/fs/ceph/quota.c b/fs/ceph/quota.c index a338a3ec0dc4..e9f7ca18cdb7 100644 --- a/fs/ceph/quota.c +++ b/fs/ceph/quota.c @@ -195,9 +195,9 @@ void ceph_cleanup_quotarealms_inodes(struct ceph_mds_client *mdsc) /* * This function walks through the snaprealm for an inode and returns the - * ceph_snap_realm for the first snaprealm that has quotas set (either max_files - * or max_bytes). If the root is reached, return the root ceph_snap_realm - * instead. + * ceph_snap_realm for the first snaprealm that has quotas set (max_files, + * max_bytes, or any, depending on the 'which_quota' argument). If the root is + * reached, return the root ceph_snap_realm instead. * * Note that the caller is responsible for calling ceph_put_snap_realm() on the * returned realm. @@ -209,7 +209,9 @@ void ceph_cleanup_quotarealms_inodes(struct ceph_mds_client *mdsc) * will be restarted. */ static struct ceph_snap_realm *get_quota_realm(struct ceph_mds_client *mdsc, - struct inode *inode, bool retry) + struct inode *inode, + enum quota_get_realm which_quota, + bool retry) { struct ceph_inode_info *ci = NULL; struct ceph_snap_realm *realm, *next; @@ -248,7 +250,7 @@ static struct ceph_snap_realm *get_quota_realm(struct ceph_mds_client *mdsc, } ci = ceph_inode(in); - has_quota = __ceph_has_any_quota(ci); + has_quota = __ceph_has_any_quota(ci, which_quota); iput(in); next = realm->parent; @@ -279,8 +281,8 @@ bool ceph_quota_is_same_realm(struct inode *old, struct inode *new) * dropped and we can then restart the whole operation. */ down_read(&mdsc->snap_rwsem); - old_realm = get_quota_realm(mdsc, old, true); - new_realm = get_quota_realm(mdsc, new, false); + old_realm = get_quota_realm(mdsc, old, QUOTA_GET_ANY, true); + new_realm = get_quota_realm(mdsc, new, QUOTA_GET_ANY, false); if (PTR_ERR(new_realm) == -EAGAIN) { up_read(&mdsc->snap_rwsem); if (old_realm) @@ -483,7 +485,8 @@ bool ceph_quota_update_statfs(struct ceph_fs_client *fsc, struct kstatfs *buf) bool is_updated = false; down_read(&mdsc->snap_rwsem); - realm = get_quota_realm(mdsc, d_inode(fsc->sb->s_root), true); + realm = get_quota_realm(mdsc, d_inode(fsc->sb->s_root), + QUOTA_GET_MAX_BYTES, true); up_read(&mdsc->snap_rwsem); if (!realm) return false; diff --git a/fs/ceph/super.h b/fs/ceph/super.h index a2e1c83ab29a..3cd96720f14a 100644 --- a/fs/ceph/super.h +++ b/fs/ceph/super.h @@ -1317,9 +1317,29 @@ extern void ceph_fs_debugfs_init(struct ceph_fs_client *client); extern void ceph_fs_debugfs_cleanup(struct ceph_fs_client *client); /* quota.c */ -static inline bool __ceph_has_any_quota(struct ceph_inode_info *ci) + +enum quota_get_realm { + QUOTA_GET_MAX_FILES, + QUOTA_GET_MAX_BYTES, + QUOTA_GET_ANY +}; + +static inline bool __ceph_has_any_quota(struct ceph_inode_info *ci, + enum quota_get_realm which) { - return ci->i_max_files || ci->i_max_bytes; + bool has_quota = false; + + switch (which) { + case QUOTA_GET_MAX_BYTES: + has_quota = !!ci->i_max_bytes; + break; + case QUOTA_GET_MAX_FILES: + has_quota = !!ci->i_max_files; + break; + default: + has_quota = !!(ci->i_max_files || ci->i_max_bytes); + } + return has_quota; } extern void ceph_adjust_quota_realms_count(struct inode *inode, bool inc); @@ -1328,10 +1348,10 @@ static inline void __ceph_update_quota(struct ceph_inode_info *ci, u64 max_bytes, u64 max_files) { bool had_quota, has_quota; - had_quota = __ceph_has_any_quota(ci); + had_quota = __ceph_has_any_quota(ci, QUOTA_GET_ANY); ci->i_max_bytes = max_bytes; ci->i_max_files = max_files; - has_quota = __ceph_has_any_quota(ci); + has_quota = __ceph_has_any_quota(ci, QUOTA_GET_ANY); if (had_quota != has_quota) ceph_adjust_quota_realms_count(&ci->vfs_inode, has_quota);