Received: by 2002:a05:6a10:5bc5:0:0:0:0 with SMTP id os5csp2187222pxb; Tue, 12 Oct 2021 00:55:32 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzCpzfBFsYdzmsFLC8GMassGRf9XdSxBnQMg7I334/RfKX48rVyAiW52ajpZ55V+Rpb7Zuw X-Received: by 2002:a05:6a00:2410:b0:40e:7d8:ad7f with SMTP id z16-20020a056a00241000b0040e07d8ad7fmr30642163pfh.25.1634025332728; Tue, 12 Oct 2021 00:55:32 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1634025332; cv=none; d=google.com; s=arc-20160816; b=yDN9UfpgD71yPGdJejPET2oWNqWUs64tqnOca/ApFSV78tkcvse9gmx5lU/Pg+JQp0 VnB7ZftjXS4aLFMAnW7NI+Pb1XvdkpfdJ6DnE1E5RMGlAaOXVILESEQ5DsnOQy9U3/2R oHhWgi3EV0KpLtTSTe3LYhzcqzrOT2Cb/a61RCYPsjckT52JglltWNLDQE4hLuYsunX3 BI4rWy/LP5zS2UaX3jmeA0WILDcirFaBhzBE+NmU10bKbHtfI/ixP5Ce4x/dwoyvHJYN yb43BzOSetNOCSxbdz/+zrxmfhz1IoStvUYIFkGrg1uDeUVTDvmkcvNXGyBf6p1ZrpFr VspQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-transfer-encoding :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=mjT+5D+rSesM57q1luFF4COIlGJ85LuX09urrK23PAw=; b=GtS1O9Ihgz5d2no2JHcdS0+0fnF7oeMmVMUHujhReAKl92/IqBdkv5FFU3iwrhH6sA /4KkcyFo0QtLKm2oQPH5jvMMVSXLRI06cwwwHXMWeljieMFQq53fLn35OSGWYOwsLUsC QAEKDaI1RAdr5jTDKGLDDNoOGUKBl7k32xG8Mc/cJFvlditVUfa5nKQqSFcgKA4yriTE Z/IGYbp5vqvI/gCQO9asfU7b7AGEm4ncpmQ8yCz4BTOXwfsLO7ywN2JEHkKN1OJ+B/ko BqIEbkCjpHHe4xEIlX3Z2B3/46fA4xF9/SBdc+L+8Pf+TjWfitjUtwK3XTifonkb/x/4 1N/g== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id w20si14903307plq.328.2021.10.12.00.55.20; Tue, 12 Oct 2021 00:55:32 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234299AbhJLH4U (ORCPT + 99 others); Tue, 12 Oct 2021 03:56:20 -0400 Received: from mga09.intel.com ([134.134.136.24]:30113 "EHLO mga09.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232541AbhJLH4U (ORCPT ); Tue, 12 Oct 2021 03:56:20 -0400 X-IronPort-AV: E=McAfee;i="6200,9189,10134"; a="226970774" X-IronPort-AV: E=Sophos;i="5.85,367,1624345200"; d="scan'208";a="226970774" Received: from orsmga006.jf.intel.com ([10.7.209.51]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 12 Oct 2021 00:54:18 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.85,367,1624345200"; d="scan'208";a="441129553" Received: from yilunxu-optiplex-7050.sh.intel.com (HELO localhost) ([10.239.159.162]) by orsmga006.jf.intel.com with ESMTP; 12 Oct 2021 00:54:15 -0700 Date: Tue, 12 Oct 2021 15:47:52 +0800 From: Xu Yilun To: Russ Weight Cc: Tom Rix , mdf@kernel.org, linux-fpga@vger.kernel.org, linux-kernel@vger.kernel.org, lgoncalv@redhat.com, hao.wu@intel.com, matthew.gerlach@intel.com Subject: Re: [PATCH v17 0/5] FPGA Image Load (previously Security Manager) Message-ID: <20211012074752.GB95330@yilunxu-OptiPlex-7050> References: <20210929230025.68961-1-russell.h.weight@intel.com> <20211009080859.GA85181@yilunxu-OptiPlex-7050> <450ed897-f726-9671-26b7-2a24bb046e89@redhat.com> <20211011014154.GA82360@yilunxu-OptiPlex-7050> <79350773-3629-2734-21c0-0314a762e722@redhat.com> <336e4827-b09a-e1ab-b67d-d8755012d71c@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <336e4827-b09a-e1ab-b67d-d8755012d71c@intel.com> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Oct 11, 2021 at 06:00:16PM -0700, Russ Weight wrote: > > > On 10/11/21 5:35 AM, Tom Rix wrote: > > > > On 10/10/21 6:41 PM, Xu Yilun wrote: > >> On Sat, Oct 09, 2021 at 05:11:20AM -0700, Tom Rix wrote: > >>> On 10/9/21 1:08 AM, Xu Yilun wrote: > >>>> On Wed, Sep 29, 2021 at 04:00:20PM -0700, Russ Weight wrote: > >>>>> The FPGA Image Load framework provides an API to upload image > >>>>> files to an FPGA device. Image files are self-describing. They could > >>>>> contain FPGA images, BMC images, Root Entry Hashes, or other device > >>>>> specific files. It is up to the lower-level device driver and the > >>>>> target device to authenticate and disposition the file data. > >>>> I've reconsider the FPGA persistent image update again, and think we > >>>> may include it in FPGA manager framework. > >>>> > >>>> Sorry I raised this topic again when it is already at patch v17, but now > >>>> I need to consider more seriously than before. > >>>> > >>>> We have consensus the FPGA persistent image update is just like a normal > >>>> firmware update which finally writes the nvmem like flash or eeprom, > >>>> while the current FPGA manager deals with the active FPGA region update > >>>> and re-activation. Could we just expand the FPGA manager and let it handle > >>>> the nvmem update as well? Many FPGA cards have nvmem and downloaders > >>>> supports updating both FPGA region and nvmem. > The fpga-image-load driver is actually just a data transfer. The class IMHO, The fpga-mgr dev is also a data transfer. The fpga-region dev is acting as the FPGA region admin responsible for gating, transfering and re-enumerating. So my opinion is to add a new data transfer type and keep a unified process. > driver has no knowledge about what the data is or where/if the data will > be stored. The fpga-image-load driver knows the data will be stored in nvmem, while the fpga-mgr knows the data will be stored in FPGA cells. They may need to know the exact physical position to store, may not, depends on what the HW engines are. > > This functionality could, of course, be merged into the fpga-mgr. I did > a proof of concept of this a while back and we discussed the pros and cons. > See this email for a recap: > > https://marc.info/?l=linux-fpga&m=161998085507374&w=2 > > Things have changed some with the evolution of the driver. The IOCTL > approach probably fits better than the sysfs implementation. At the time > it seemed that a merge would add unnecessary complexity without adding value. I think at least developers don't have to go through 2 sets of software stacks which are of the same concept. And adding some new features like optionally threading or canceling data transfer are also good to FPGA region update. And the nvmem update could also be benifit from exsiting implementations like scatter-gather buffers, in-kernel firmware loading. I try to explain myself according to each of your concern from that mail thread: Purpose of the 2 updates ======================== As I said before, I think they are both data transfer devices. FPGA region update gets extra support from fpga-region & fpga-bridge, and FPGA nvmem update could be a standalone fpga-mgr. Extra APIs that are unique to nvmem update ========================================== cdev APIs for nvmem update: Yes we need to expand the functionality so we need them. available_images, image_load APIs for loading nvmem content to FPGA region: These are features in later patchsets which are not submitted, but we could talk about it in advance. I think this is actually a FPGA region update from nvmem, it also requires gating, data loading (no SW transfer) and re-enumeration, or a single command to image_load HW may result system disordered. The FPGA framework now only supports update from in-kernel user data, maybe we add support for update from nvmem later. fpga-mgr state extend: I think it could be extended, The current states are not perfect, adding something like IDLE or READY is just fine. fpga-mgr status extend: Add general error definitions as needed. If there is some status that is quite vendor specific, expose it in low-level driver. remaining-size: Nice to have for all. Threading the update ==================== Also a good option for FPGA region update, maybe we also have a slow FPGA reprogrammer? Cancelling the update ==================== Also a good option for FPGA region update if HW engine supports. Thanks, Yilun > > >>>> > >>>> According to the patchset, the basic workflow of the 2 update types are > >>>> quite similar, get the data, prepare for the HW, write and complete. > >>>> They are already implemented in FPGA manager. We've discussed some > >>>> differences like threading or canceling the update, which are > >>>> not provided by FPGA manager but they may also nice to have for FPGA > >>>> region update. An FPGA region update may also last for a long time?? > >>>> So I think having 2 sets of similar frameworks in FPGA is unnecessary. > >>>> > >>>> My quick mind is that we add some flags in struct fpga_mgr & struct > >>>> fpga_image_info to indicate the HW capability (support FPGA region > >>>> update or nvmem update or both) of the download engine and the provided > >>>> image type. Then the low-level driver knows how to download if it > >>>> supports both image types.An char device could be added for each fpga manager dev, providing the > >>>> user APIs for nvmem update. We may not use the char dev for FPGA region > >>>> update cause it changes the system HW devices and needs device hotplug > >>>> in FPGA region. We'd better leave it to FPGA region class, this is > >>>> another topic. > I'll give this some more thought and see if I can come up with some RFC > patches. > > - Russ > >>>> > >>>> More discussion is appreciated. > >>> I also think fpga_mgr could be extended. > >>> > >>> In this patchset, > >>> > >>> https://lore.kernel.org/linux-fpga/20210625195849.837976-1-trix@redhat.com/ > >>> > >>> A second, similar set of write ops was added to fpga_manger_ops, > >>> > >>> new bit/flag was added to fpga_image_info > >>> > >>> The intent was for dfl to add their specific ops to cover what is done in > >>> this patchset. > >> I think we don't have to add 2 ops for reconfig & reimage in framework, > >> the 2 processes are almost the same. > >> > >> Just add the _REIMAGE (or something else, NVMEM?) flag for > >> fpga_image_info, and low level drivers handle it as they do for other > >> flags. > >> > >> How do you think? > > > > A single set is fine. > > > > A difficult part of is the length of? time to do the write. The existing write should be improved to use a worker thread. > > > > Tom > > > >> > >> Thanks, > >> Yilun > >> > >>> Any other driver would do similar. > >>> > >>> Is this close to what you are thinking ? > >>> > >>> Tom > >>> > >>>> Thanks, > >>>> Yilun > >>>> > >