Received: by 2002:a25:c593:0:0:0:0:0 with SMTP id v141csp4834460ybe; Mon, 9 Sep 2019 15:38:48 -0700 (PDT) X-Google-Smtp-Source: APXvYqyQeOtKAkpf5c7Tlzx+vkxIgG3FXZuT4aq70zF6E1WTXyd4M7F7Gx+GXX5XlplHHNSZcScy X-Received: by 2002:a05:6402:2d0:: with SMTP id b16mr27955003edx.147.1568068728139; Mon, 09 Sep 2019 15:38:48 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1568068728; cv=none; d=google.com; s=arc-20160816; b=F9dOl36lE1Zf6on0sZrBzYUyGrwxpOXHiwOP0F+V9GyB7YKTrrGV3jMslLRJr3x6ee 9wZY0D8W/4LkwOZjWE1+b82IZDWAflWPaBZmrMSzCt3lbFpPdb8p38NASWL5V/FYH61i 8ZLWfqTdaoMsfEKle3bi+hmh3epaV0iZEq5mBVY+siNiIhoQis78dHzC+eIdanOftQ+W uMWrKv7vwV1rFSnuXIGvCEcr4dwQRkJfELPZz0XhWcbH/UecRDxwoOf1857C0YVO6DA0 liHcCsr9E3tOGFXlWVepCz9y/7C6g6E2YckrnKWaKCaMo2a+a+ybGR4Co58AgP7+LdhV NbdQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=e1/cIJxx/RW62uvfTm+vZSA/WqpE9yddHSNRvrze3Wg=; b=mONsUZZM/YZZyFF6Mibb9hGfW8y8B3op6pjHx025Ri56kgjHKr8XxNXdp4aoKESee2 a5YxrVnVTbndqy6ucAZSYEtBWcpyi+FrsmEFymIVuvsmC7FkzGUQELiK40BOOtnXvsRb E3YdM1p7xnVK05uSwUHBRfo3pfLLADk4g8/AbWx3TTl1DRNEo5jaRpS0zPwDRvN04GGv 5N6Z9BjqXBrsIuIJhFZM6z49zOa6HZwy7yjuMvoUey5AHiWEF0BS7fIjzwg11UQ8oowi D7ZAPZmeFmeItD+5NwCis1Pv43hs8KcQIJSpokF1yVfUJlY/iH10yxnDLuGzBRR+wDg2 ruUw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id qh7si17979ejb.280.2019.09.09.15.38.23; Mon, 09 Sep 2019 15:38:48 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730486AbfIIK3q (ORCPT + 99 others); Mon, 9 Sep 2019 06:29:46 -0400 Received: from mx2.suse.de ([195.135.220.15]:38452 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1730369AbfIIK3p (ORCPT ); Mon, 9 Sep 2019 06:29:45 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id 512DBAE86; Mon, 9 Sep 2019 10:28:36 +0000 (UTC) From: Luis Henriques To: Jeff Layton , Sage Weil , Ilya Dryomov Cc: ceph-devel@vger.kernel.org, linux-kernel@vger.kernel.org, Luis Henriques Subject: [PATCH v2] ceph: allow object copies across different filesystems in the same cluster Date: Mon, 9 Sep 2019 11:28:34 +0100 Message-Id: <20190909102834.16246-1-lhenriques@suse.com> In-Reply-To: <87k1ahojri.fsf@suse.com> References: <87k1ahojri.fsf@suse.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org OSDs are able to perform object copies across different pools. Thus, there's no need to prevent copy_file_range from doing remote copies if the source and destination superblocks are different. Only return -EXDEV if they have different fsid (the cluster ID). Signed-off-by: Luis Henriques --- fs/ceph/file.c | 18 ++++++++++++++---- 1 file changed, 14 insertions(+), 4 deletions(-) Hi, Here's the patch changelog since initial submittion: - Dropped have_fsid checks on client structs - Use %pU to print the fsid instead of raw hex strings (%*ph) - Fixed 'To:' field in email so that this time the patch hits vger Cheers, -- Luis diff --git a/fs/ceph/file.c b/fs/ceph/file.c index 685a03cc4b77..4a624a1dd0bb 100644 --- a/fs/ceph/file.c +++ b/fs/ceph/file.c @@ -1904,6 +1904,7 @@ static ssize_t __ceph_copy_file_range(struct file *src_file, loff_t src_off, struct ceph_inode_info *src_ci = ceph_inode(src_inode); struct ceph_inode_info *dst_ci = ceph_inode(dst_inode); struct ceph_cap_flush *prealloc_cf; + struct ceph_fs_client *src_fsc = ceph_inode_to_client(src_inode); struct ceph_object_locator src_oloc, dst_oloc; struct ceph_object_id src_oid, dst_oid; loff_t endoff = 0, size; @@ -1915,8 +1916,17 @@ static ssize_t __ceph_copy_file_range(struct file *src_file, loff_t src_off, if (src_inode == dst_inode) return -EINVAL; - if (src_inode->i_sb != dst_inode->i_sb) - return -EXDEV; + if (src_inode->i_sb != dst_inode->i_sb) { + struct ceph_fs_client *dst_fsc = ceph_inode_to_client(dst_inode); + + if (ceph_fsid_compare(&src_fsc->client->fsid, + &dst_fsc->client->fsid)) { + dout("Copying object across different clusters:"); + dout(" src fsid: %pU dst fsid: %pU\n", + &src_fsc->client->fsid, &dst_fsc->client->fsid); + return -EXDEV; + } + } if (ceph_snap(dst_inode) != CEPH_NOSNAP) return -EROFS; @@ -1928,7 +1938,7 @@ static ssize_t __ceph_copy_file_range(struct file *src_file, loff_t src_off, * efficient). */ - if (ceph_test_mount_opt(ceph_inode_to_client(src_inode), NOCOPYFROM)) + if (ceph_test_mount_opt(src_fsc, NOCOPYFROM)) return -EOPNOTSUPP; if ((src_ci->i_layout.stripe_unit != dst_ci->i_layout.stripe_unit) || @@ -2044,7 +2054,7 @@ static ssize_t __ceph_copy_file_range(struct file *src_file, loff_t src_off, dst_ci->i_vino.ino, dst_objnum); /* Do an object remote copy */ err = ceph_osdc_copy_from( - &ceph_inode_to_client(src_inode)->client->osdc, + &src_fsc->client->osdc, src_ci->i_vino.snap, 0, &src_oid, &src_oloc, CEPH_OSD_OP_FLAG_FADVISE_SEQUENTIAL |