Received: by 2002:ac0:aed5:0:0:0:0:0 with SMTP id t21csp656503imb; Fri, 1 Mar 2019 10:20:12 -0800 (PST) X-Google-Smtp-Source: APXvYqwSble96xVMCeIx6Go5wWU2/JGGDEpQrrsxSrCaAxZx45QN+pDBo/qPVeji/NJ+oIcYuKts X-Received: by 2002:a65:5142:: with SMTP id g2mr6095968pgq.149.1551464412354; Fri, 01 Mar 2019 10:20:12 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1551464412; cv=none; d=google.com; s=arc-20160816; b=mSZRyRntc4AYiOTFfutXmlSJtZ3+Tc4vZSYIF2r/y6iJUL/qlfpwqPQ1pz8YJlx3dN mHvPUjmNzV6/sUZFoNF5Ky6XatkRxwpDX/Fc1WR/zHCWn3denxbpraDgTDxZlhJZykWW FjZvWagG5iNCDsP+U4XuEjIxOZn8sU9mHti35bVvD0wCq/RZtG6dNAJuJAi9jWNcunX8 X0VSH09byP99ej3RmFZF/Y2SviGf6KtQbGntElRRhfvTvYNyHKcdT3CMIvUH1EdV/H+M HSEkO9ZVO0hVzlRmWqNmivoYfhZM7mkGhN6UMzKad6kMvaQsOGQeHKBkCF4lNtLPUXtZ 9yrw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=qBd4tQGqqysLcGP64q662pIj0HWYWVsWnfcPgbt94EA=; b=qLicj0lfuHcQeyFNEfvnPy2mh3ItY3/bX+Our+d0YHX4imAYCU7hhfPI2Y4AYt2RhR iheJsmzJVXsVDQdZcZ+cRfiBYzfKCXPuCHGhs5D3MP65Jt5CupG24QmLBN9a/UA1bUWx czCtE9vcjrkBzpSZcWwm7Jcgb4osLX5/e1w0nw4RFDq9TDMyIrsm4YabsR6vSjPppHxd 56kyITsbTlMRQkP24qjK+a1n3HxIRg25mXgVShWf01YEVP+az3tt5s84oxglo0lxQh4J XFQjT2EgR3/OwF2NMRTc76k7x/vcuRG2lu+uc2pCI3VWKMg3Y1Y7pxEiLO1KlxWif82i dazA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id i69si10721870pge.103.2019.03.01.10.19.56; Fri, 01 Mar 2019 10:20:12 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2389748AbfCAR6B (ORCPT + 99 others); Fri, 1 Mar 2019 12:58:01 -0500 Received: from mx2.suse.de ([195.135.220.15]:43330 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S2389722AbfCAR6B (ORCPT ); Fri, 1 Mar 2019 12:58:01 -0500 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id E46C3AED7; Fri, 1 Mar 2019 17:57:58 +0000 (UTC) From: Luis Henriques To: "Yan, Zheng" , Sage Weil , Ilya Dryomov Cc: ceph-devel@vger.kernel.org, linux-kernel@vger.kernel.org, Luis Henriques , Hendrik Peyerl Subject: [RFC PATCH 2/2] ceph: quota: fix quota subdir mounts Date: Fri, 1 Mar 2019 17:57:52 +0000 Message-Id: <20190301175752.17808-3-lhenriques@suse.com> In-Reply-To: <20190301175752.17808-1-lhenriques@suse.com> References: <20190301175752.17808-1-lhenriques@suse.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org The CephFS kernel client doesn't enforce quotas that are set in a directory that isn't visible in the mount point. For example, given the path '/dir1/dir2', if quotas are set in 'dir1' and the mount is done in with mount -t ceph ::/dir1/ /mnt then the client can't access the 'dir1' inode from the quota realm dir2 belongs to. This patch fixes this by simply doing an MDS LOOKUPINO Op and grabbing a reference to it (so that it doesn't disappear again). This also requires an extra field in ceph_snap_realm so that we know we have to release that reference when destroying the realm. Links: https://tracker.ceph.com/issues/3848 Reported-by: Hendrik Peyerl Signed-off-by: Luis Henriques --- fs/ceph/caps.c | 2 +- fs/ceph/quota.c | 30 +++++++++++++++++++++++++++--- fs/ceph/snap.c | 3 +++ fs/ceph/super.h | 2 ++ 4 files changed, 33 insertions(+), 4 deletions(-) diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c index bba28a5034ba..e79994ff53d6 100644 --- a/fs/ceph/caps.c +++ b/fs/ceph/caps.c @@ -1035,7 +1035,7 @@ static void drop_inode_snap_realm(struct ceph_inode_info *ci) list_del_init(&ci->i_snap_realm_item); ci->i_snap_realm_counter++; ci->i_snap_realm = NULL; - if (realm->ino == ci->i_vino.ino) + if ((realm->ino == ci->i_vino.ino) && !realm->own_inode) realm->inode = NULL; spin_unlock(&realm->inodes_with_caps_lock); ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc, diff --git a/fs/ceph/quota.c b/fs/ceph/quota.c index 9455d3aef0c3..f6b972d222e4 100644 --- a/fs/ceph/quota.c +++ b/fs/ceph/quota.c @@ -22,7 +22,16 @@ void ceph_adjust_quota_realms_count(struct inode *inode, bool inc) static inline bool ceph_has_realms_with_quotas(struct inode *inode) { struct ceph_mds_client *mdsc = ceph_inode_to_client(inode)->mdsc; - return atomic64_read(&mdsc->quotarealms_count) > 0; + struct super_block *sb = mdsc->fsc->sb; + + if (atomic64_read(&mdsc->quotarealms_count) > 0) + return true; + /* if root is the real CephFS root, we don't have quota realms */ + if (sb->s_root->d_inode && + (sb->s_root->d_inode->i_ino == CEPH_INO_ROOT)) + return false; + /* otherwise, we can't know for sure */ + return true; } void ceph_handle_quota(struct ceph_mds_client *mdsc, @@ -166,6 +175,7 @@ static bool check_quota_exceeded(struct inode *inode, enum quota_check_op op, return false; down_read(&mdsc->snap_rwsem); +restart: realm = ceph_inode(inode)->i_snap_realm; if (realm) ceph_get_snap_realm(mdsc, realm); @@ -176,8 +186,22 @@ static bool check_quota_exceeded(struct inode *inode, enum quota_check_op op, spin_lock(&realm->inodes_with_caps_lock); in = realm->inode ? igrab(realm->inode) : NULL; spin_unlock(&realm->inodes_with_caps_lock); - if (!in) - break; + if (!in) { + up_read(&mdsc->snap_rwsem); + in = ceph_lookup_inode(inode->i_sb, realm->ino); + down_read(&mdsc->snap_rwsem); + if (IS_ERR(in)) { + pr_warn("Can't lookup inode %llx (err: %ld)\n", + realm->ino, PTR_ERR(in)); + break; + } + spin_lock(&realm->inodes_with_caps_lock); + realm->inode = in; + realm->own_inode = true; + spin_unlock(&realm->inodes_with_caps_lock); + ceph_put_snap_realm(mdsc, realm); + goto restart; + } ci = ceph_inode(in); spin_lock(&ci->i_ceph_lock); diff --git a/fs/ceph/snap.c b/fs/ceph/snap.c index f74193da0e09..c84ed8e8526a 100644 --- a/fs/ceph/snap.c +++ b/fs/ceph/snap.c @@ -117,6 +117,7 @@ static struct ceph_snap_realm *ceph_create_snap_realm( atomic_set(&realm->nref, 1); /* for caller */ realm->ino = ino; + realm->own_inode = false; INIT_LIST_HEAD(&realm->children); INIT_LIST_HEAD(&realm->child_item); INIT_LIST_HEAD(&realm->empty_item); @@ -184,6 +185,8 @@ static void __destroy_snap_realm(struct ceph_mds_client *mdsc, kfree(realm->prior_parent_snaps); kfree(realm->snaps); ceph_put_snap_context(realm->cached_context); + if (realm->own_inode && realm->inode) + iput(realm->inode); kfree(realm); } diff --git a/fs/ceph/super.h b/fs/ceph/super.h index ce51e98b08ec..3f0d74d2150f 100644 --- a/fs/ceph/super.h +++ b/fs/ceph/super.h @@ -764,6 +764,8 @@ struct ceph_snap_realm { atomic_t nref; struct rb_node node; + bool own_inode; /* true if we hold a ref to the inode */ + u64 created, seq; u64 parent_ino; u64 parent_since; /* snapid when our current parent became so */