Received: by 2002:ac0:a5a7:0:0:0:0:0 with SMTP id m36-v6csp1204514imm; Fri, 13 Jul 2018 13:37:54 -0700 (PDT) X-Google-Smtp-Source: AAOMgpeBkvA9AQqbs1X5wgvRvFTaC4DTopBaUB2A6leaRNU4MJap1amapsUdcrWmqL7xlwMmqDbn X-Received: by 2002:a17:902:b594:: with SMTP id a20-v6mr2525904pls.140.1531514274210; Fri, 13 Jul 2018 13:37:54 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1531514274; cv=none; d=google.com; s=arc-20160816; b=s8eUmw9tdS3eRwTX23Ug6oZNbmpQNsqrWFVM3io6Yp+JQOej22OJG2AR+RfYy3+ALI r7NohxN+A/M+sSYACYO7MO36hslHMHJAByPwyimHU4u4o2dRpfd5+MEXdqpd4TYKyxB6 hitt6nm9OodxpZa8QAkr+EEvcShqAWR9Pe33HT6zpua+5zsE1uca1TaSCMGSTlpOJglh Tln25fn87LG8/YTypcnYdeApi/Q70liDn32PA05NPQ+dsK3H9iB/OMtQegqy80yCzYsQ JPON1JLwKOQc3XRjXFJcVs/vzaKwXVaGgzHK9p7UV8OKJ+UfzRcBolDghZpOWcyw6Ihb UB2Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:subject:cc:to:from:date :arc-authentication-results; bh=7sQH45l5XhU2GAGOsIb0mIYSV73y3D4wrJFTPMWQYh4=; b=LyPrATAJx6bhhWFK6bE/5vC4+JxaeocGyeyTalIOuT0y0GtHkwn/UBKOGfO9v1+EsA OBvbxxmCOr3TsV0BaUxikGy0jWQDke7KC6t+0AIJp/v+mQW7VIwzFdRTy0bbMecyhenH lOE8zaE8NGwaWDHqVWCme8xddr95S24Jv90weC8ofsv3H6GsaOaUp/kEXR6pIYHoN8Ei TvFJrzg0qTC1WLl4LmWXRWycDs/vrSktJ+N+ySKu87Y4zeQhadoe0vRnYTvsbFg8iWlB Rm1LwkpjBgx5zpdokatNreSEHH3Uueuftz9Qc8s6WQr0PTwci4kWm9VJmEizfQC00Frg +8eA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 31-v6si23949314plk.191.2018.07.13.13.37.39; Fri, 13 Jul 2018 13:37:54 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1731964AbeGMUw1 (ORCPT + 99 others); Fri, 13 Jul 2018 16:52:27 -0400 Received: from mx3-rdu2.redhat.com ([66.187.233.73]:52788 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1731883AbeGMUw1 (ORCPT ); Fri, 13 Jul 2018 16:52:27 -0400 Received: from smtp.corp.redhat.com (int-mx03.intmail.prod.int.rdu2.redhat.com [10.11.54.3]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 7009E40122CF; Fri, 13 Jul 2018 20:36:13 +0000 (UTC) Received: from doriath (ovpn-117-239.phx2.redhat.com [10.3.117.239]) by smtp.corp.redhat.com (Postfix) with ESMTP id 7F824112C268; Fri, 13 Jul 2018 20:36:00 +0000 (UTC) Date: Fri, 13 Jul 2018 16:35:59 -0400 From: Luiz Capitulino To: Pankaj Gupta Cc: linux-kernel@vger.kernel.org, kvm@vger.kernel.org, qemu-devel@nongnu.org, linux-nvdimm@ml01.01.org, jack@suse.cz, stefanha@redhat.com, dan.j.williams@intel.com, riel@surriel.com, haozhong.zhang@intel.com, nilal@redhat.com, kwolf@redhat.com, pbonzini@redhat.com, ross.zwisler@intel.com, david@redhat.com, xiaoguangrong.eric@gmail.com, hch@infradead.org, mst@redhat.com, niteshnarayanlal@hotmail.com, imammedo@redhat.com, eblake@redhat.com Subject: Re: [RFC v3 1/2] libnvdimm: Add flush callback for virtio pmem Message-ID: <20180713163559.692aca30@doriath> In-Reply-To: <20180713075232.9575-2-pagupta@redhat.com> References: <20180713075232.9575-1-pagupta@redhat.com> <20180713075232.9575-2-pagupta@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit X-Scanned-By: MIMEDefang 2.78 on 10.11.54.3 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.5]); Fri, 13 Jul 2018 20:36:13 +0000 (UTC) X-Greylist: inspected by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.5]); Fri, 13 Jul 2018 20:36:13 +0000 (UTC) for IP:'10.11.54.3' DOMAIN:'int-mx03.intmail.prod.int.rdu2.redhat.com' HELO:'smtp.corp.redhat.com' FROM:'lcapitulino@redhat.com' RCPT:'' Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, 13 Jul 2018 13:22:30 +0530 Pankaj Gupta wrote: > This patch adds functionality to perform flush from guest to host > over VIRTIO. We are registering a callback based on 'nd_region' type. > As virtio_pmem driver requires this special flush interface, for rest > of the region types we are registering existing flush function. > Also report the error returned by virtio flush interface. This patch doesn't apply against latest upstream. A few more comments below. > > Signed-off-by: Pankaj Gupta > --- > drivers/nvdimm/nd.h | 1 + > drivers/nvdimm/pmem.c | 4 ++-- > drivers/nvdimm/region_devs.c | 24 ++++++++++++++++++------ > include/linux/libnvdimm.h | 5 ++++- > 4 files changed, 25 insertions(+), 9 deletions(-) > > diff --git a/drivers/nvdimm/nd.h b/drivers/nvdimm/nd.h > index 32e0364..1b62f79 100644 > --- a/drivers/nvdimm/nd.h > +++ b/drivers/nvdimm/nd.h > @@ -159,6 +159,7 @@ struct nd_region { > struct badblocks bb; > struct nd_interleave_set *nd_set; > struct nd_percpu_lane __percpu *lane; > + int (*flush)(struct device *dev); > struct nd_mapping mapping[0]; > }; > > diff --git a/drivers/nvdimm/pmem.c b/drivers/nvdimm/pmem.c > index 9d71492..29fd2cd 100644 > --- a/drivers/nvdimm/pmem.c > +++ b/drivers/nvdimm/pmem.c > @@ -180,7 +180,7 @@ static blk_qc_t pmem_make_request(struct request_queue *q, struct bio *bio) > struct nd_region *nd_region = to_region(pmem); > > if (bio->bi_opf & REQ_FLUSH) > - nvdimm_flush(nd_region); > + bio->bi_status = nvdimm_flush(nd_region); > > do_acct = nd_iostat_start(bio, &start); > bio_for_each_segment(bvec, bio, iter) { > @@ -196,7 +196,7 @@ static blk_qc_t pmem_make_request(struct request_queue *q, struct bio *bio) > nd_iostat_end(bio, start); > > if (bio->bi_opf & REQ_FUA) > - nvdimm_flush(nd_region); > + bio->bi_status = nvdimm_flush(nd_region); > > bio_endio(bio); > return BLK_QC_T_NONE; > diff --git a/drivers/nvdimm/region_devs.c b/drivers/nvdimm/region_devs.c > index a612be6..124aae7 100644 > --- a/drivers/nvdimm/region_devs.c > +++ b/drivers/nvdimm/region_devs.c > @@ -1025,6 +1025,7 @@ static struct nd_region *nd_region_create(struct nvdimm_bus *nvdimm_bus, > dev->of_node = ndr_desc->of_node; > nd_region->ndr_size = resource_size(ndr_desc->res); > nd_region->ndr_start = ndr_desc->res->start; > + nd_region->flush = ndr_desc->flush; > nd_device_register(dev); > > return nd_region; > @@ -1065,13 +1066,10 @@ struct nd_region *nvdimm_volatile_region_create(struct nvdimm_bus *nvdimm_bus, > } > EXPORT_SYMBOL_GPL(nvdimm_volatile_region_create); > > -/** > - * nvdimm_flush - flush any posted write queues between the cpu and pmem media > - * @nd_region: blk or interleaved pmem region > - */ > -void nvdimm_flush(struct nd_region *nd_region) > +void pmem_flush(struct device *dev) > { > - struct nd_region_data *ndrd = dev_get_drvdata(&nd_region->dev); > + struct nd_region_data *ndrd = dev_get_drvdata(dev); > + struct nd_region *nd_region = to_nd_region(dev); > int i, idx; > > /* > @@ -1094,6 +1092,20 @@ void nvdimm_flush(struct nd_region *nd_region) > writeq(1, ndrd_get_flush_wpq(ndrd, i, idx)); > wmb(); > } > + > +/** > + * nvdimm_flush - flush any posted write queues between the cpu and pmem media > + * @nd_region: blk or interleaved pmem region > + */ > +int nvdimm_flush(struct nd_region *nd_region) > +{ > + if (nd_region->flush) > + return(nd_region->flush(&nd_region->dev)); > + > + pmem_flush(&nd_region->dev); IMHO, a better way of doing this would be to allow nvdimm_flush() to be overridden. That is, in nd_region_create() you set nd_region->flush to the original nvdimm_flush() if ndr_desc->flush is NULL. And then always call nd_region->flush() where nvdimm_flush() is called today. > + > + return 0; > +} > EXPORT_SYMBOL_GPL(nvdimm_flush); > > /** > diff --git a/include/linux/libnvdimm.h b/include/linux/libnvdimm.h > index 097072c..33b617f 100644 > --- a/include/linux/libnvdimm.h > +++ b/include/linux/libnvdimm.h > @@ -126,6 +126,7 @@ struct nd_region_desc { > int numa_node; > unsigned long flags; > struct device_node *of_node; > + int (*flush)(struct device *dev); > }; > > struct device; > @@ -201,7 +202,9 @@ unsigned long nd_blk_memremap_flags(struct nd_blk_region *ndbr); > unsigned int nd_region_acquire_lane(struct nd_region *nd_region); > void nd_region_release_lane(struct nd_region *nd_region, unsigned int lane); > u64 nd_fletcher64(void *addr, size_t len, bool le); > -void nvdimm_flush(struct nd_region *nd_region); > +int nvdimm_flush(struct nd_region *nd_region); > +void pmem_set_flush(struct nd_region *nd_region, void (*flush) > + (struct device *)); It seems pmem_set_flush() doesn't exist. > int nvdimm_has_flush(struct nd_region *nd_region); > int nvdimm_has_cache(struct nd_region *nd_region); >