Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp52712pxk; Tue, 15 Sep 2020 20:12:52 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxVpxYc6/IT12KuJuIfQjb2bTlbEK6/USPVyCuQLxPYIRL47isYZEvfECHZtv+qBsw0zh9g X-Received: by 2002:a05:6402:2c3:: with SMTP id b3mr25522420edx.213.1600225972607; Tue, 15 Sep 2020 20:12:52 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1600225972; cv=none; d=google.com; s=arc-20160816; b=wZ6jQL8wxzIk2zKK4iPY6N76GOkObfGWfnFtJrgoX548FmFXFd47n7XnCCBCRbxCUV vTQNOyxrYQX6/eMmSiptEgp6k3wkjSIGYJpK4jIvcxfn+9wnJ6Vr0nCK5f2AoIyWR8eX URD5KK0Pvsf+OC372LlK//Kod7jwC3xCBrDI9R3obK37R1TsXEDdR8YnZQVk8YdlFcbv 9M4zFc1cSNRDcuQ3jfKAac9+C6DGYBS7wklhgSh70dshC9W2+00pTrEHT+okCUSQpuXm k9KuP7p9stwoJQByNE8RpqUjI+PcqqqCiq5vUQeaHpYgpucm5MmrHS/JLeffCc2H2R00 xiIA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-language :content-transfer-encoding:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:dkim-signature; bh=pUfKtm728JV5HfEPBrNg85FNcQ7Ve2VTddaezZoQ5WY=; b=pLcZv9yYapS0dVyPlwn3xVH7/ZJk0v74LCRh7IULEcUgkFFcE9WQSXmZXMmrpy1aw3 YVnYqWeI2zww1YgAcs/j/4V9jbVgnJrhuu+DBffLMC0jNntlviywb6G/sOqw721tWR0d jMgfZrs3AS75qGUlIDtcFjne2/m+uhQwkHM5Cf7IPyem4eV7BZ4ehw/K43CjMoU+h2RS XbZeBDM9P0gzM0STs1Q3J4YJaHMuzJhxKmU4aSbDrgYKANuH632nz76A0vY39Ojjmxo9 8LbRHqJgsmr9lft64E/u/umiNivEvZc0CBFM57uHPKojWvCSv0clB72HEYHhRZnVvkvH vfTw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=NTFSBk7Q; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id 19si10773106ejb.381.2020.09.15.20.12.30; Tue, 15 Sep 2020 20:12:52 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b=NTFSBk7Q; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726328AbgIPDLQ (ORCPT + 99 others); Tue, 15 Sep 2020 23:11:16 -0400 Received: from us-smtp-delivery-1.mimecast.com ([207.211.31.120]:51080 "EHLO us-smtp-1.mimecast.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726303AbgIPDLP (ORCPT ); Tue, 15 Sep 2020 23:11:15 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1600225873; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=pUfKtm728JV5HfEPBrNg85FNcQ7Ve2VTddaezZoQ5WY=; b=NTFSBk7Q6S1IZx0LmnarbI8s37XwMNfgF0a242IrTudlhcOSk14kbtTCywzcvbSJwBZBYR t4seOcvTOk5QymsgfLuUugvvFzzysPVOpv4Td5laoqwll9gebSqRQz6sfc5cOmxZdBQMhp Ts1C/me0eJmH75uxkTyJkdvCi5ezGrg= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-165-J5rmrvsYOhil88kP7owOHQ-1; Tue, 15 Sep 2020 23:11:09 -0400 X-MC-Unique: J5rmrvsYOhil88kP7owOHQ-1 Received: from smtp.corp.redhat.com (int-mx05.intmail.prod.int.phx2.redhat.com [10.5.11.15]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 6EC658015FD; Wed, 16 Sep 2020 03:11:06 +0000 (UTC) Received: from [10.72.13.186] (ovpn-13-186.pek2.redhat.com [10.72.13.186]) by smtp.corp.redhat.com (Postfix) with ESMTP id C29ED75124; Wed, 16 Sep 2020 03:10:52 +0000 (UTC) Subject: Re: [RFC PATCH 00/22] Enhance VHOST to enable SoC-to-SoC communication To: Kishon Vijay Abraham I , Cornelia Huck Cc: "Michael S. Tsirkin" , Ohad Ben-Cohen , Bjorn Andersson , Jon Mason , Dave Jiang , Allen Hubbe , Lorenzo Pieralisi , Bjorn Helgaas , Paolo Bonzini , Stefan Hajnoczi , Stefano Garzarella , linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-remoteproc@vger.kernel.org, linux-ntb@googlegroups.com, linux-pci@vger.kernel.org, kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, netdev@vger.kernel.org References: <20200702082143.25259-1-kishon@ti.com> <20200702055026-mutt-send-email-mst@kernel.org> <603970f5-3289-cd53-82a9-aa62b292c552@redhat.com> <14c6cad7-9361-7fa4-e1c6-715ccc7e5f6b@ti.com> <59fd6a0b-8566-44b7-3dae-bb52b468219b@redhat.com> <45a8a97c-2061-13ee-5da8-9877a4a3b8aa@ti.com> <20200828123409.4cd2a812.cohuck@redhat.com> <9cd58cd1-0041-3d98-baf7-6e5bc2e7e317@redhat.com> <5733dbfc-76c1-45dc-6dce-ef5449eacc73@redhat.com> <181ae83d-edeb-9406-27cc-1195fe29ae95@ti.com> From: Jason Wang Message-ID: Date: Wed, 16 Sep 2020 11:10:50 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: <181ae83d-edeb-9406-27cc-1195fe29ae95@ti.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.15 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2020/9/15 下午11:47, Kishon Vijay Abraham I wrote: > Hi Jason, > > On 15/09/20 1:48 pm, Jason Wang wrote: >> Hi Kishon: >> >> On 2020/9/14 下午3:23, Kishon Vijay Abraham I wrote: >>>> Then you need something that is functional equivalent to virtio PCI >>>> which is actually the concept of vDPA (e.g vDPA provides alternatives if >>>> the queue_sel is hard in the EP implementation). >>> Okay, I just tried to compare the 'struct vdpa_config_ops' and 'struct >>> vhost_config_ops' ( introduced in [RFC PATCH 03/22] vhost: Add ops for >>> the VHOST driver to configure VHOST device). >>> >>> struct vdpa_config_ops { >>>     /* Virtqueue ops */ >>>     int (*set_vq_address)(struct vdpa_device *vdev, >>>                   u16 idx, u64 desc_area, u64 driver_area, >>>                   u64 device_area); >>>     void (*set_vq_num)(struct vdpa_device *vdev, u16 idx, u32 num); >>>     void (*kick_vq)(struct vdpa_device *vdev, u16 idx); >>>     void (*set_vq_cb)(struct vdpa_device *vdev, u16 idx, >>>               struct vdpa_callback *cb); >>>     void (*set_vq_ready)(struct vdpa_device *vdev, u16 idx, bool ready); >>>     bool (*get_vq_ready)(struct vdpa_device *vdev, u16 idx); >>>     int (*set_vq_state)(struct vdpa_device *vdev, u16 idx, >>>                 const struct vdpa_vq_state *state); >>>     int (*get_vq_state)(struct vdpa_device *vdev, u16 idx, >>>                 struct vdpa_vq_state *state); >>>     struct vdpa_notification_area >>>     (*get_vq_notification)(struct vdpa_device *vdev, u16 idx); >>>     /* vq irq is not expected to be changed once DRIVER_OK is set */ >>>     int (*get_vq_irq)(struct vdpa_device *vdv, u16 idx); >>> >>>     /* Device ops */ >>>     u32 (*get_vq_align)(struct vdpa_device *vdev); >>>     u64 (*get_features)(struct vdpa_device *vdev); >>>     int (*set_features)(struct vdpa_device *vdev, u64 features); >>>     void (*set_config_cb)(struct vdpa_device *vdev, >>>                   struct vdpa_callback *cb); >>>     u16 (*get_vq_num_max)(struct vdpa_device *vdev); >>>     u32 (*get_device_id)(struct vdpa_device *vdev); >>>     u32 (*get_vendor_id)(struct vdpa_device *vdev); >>>     u8 (*get_status)(struct vdpa_device *vdev); >>>     void (*set_status)(struct vdpa_device *vdev, u8 status); >>>     void (*get_config)(struct vdpa_device *vdev, unsigned int offset, >>>                void *buf, unsigned int len); >>>     void (*set_config)(struct vdpa_device *vdev, unsigned int offset, >>>                const void *buf, unsigned int len); >>>     u32 (*get_generation)(struct vdpa_device *vdev); >>> >>>     /* DMA ops */ >>>     int (*set_map)(struct vdpa_device *vdev, struct vhost_iotlb *iotlb); >>>     int (*dma_map)(struct vdpa_device *vdev, u64 iova, u64 size, >>>                u64 pa, u32 perm); >>>     int (*dma_unmap)(struct vdpa_device *vdev, u64 iova, u64 size); >>> >>>     /* Free device resources */ >>>     void (*free)(struct vdpa_device *vdev); >>> }; >>> >>> +struct vhost_config_ops { >>> +    int (*create_vqs)(struct vhost_dev *vdev, unsigned int nvqs, >>> +              unsigned int num_bufs, struct vhost_virtqueue *vqs[], >>> +              vhost_vq_callback_t *callbacks[], >>> +              const char * const names[]); >>> +    void (*del_vqs)(struct vhost_dev *vdev); >>> +    int (*write)(struct vhost_dev *vdev, u64 vhost_dst, void *src, >>> int len); >>> +    int (*read)(struct vhost_dev *vdev, void *dst, u64 vhost_src, int >>> len); >>> +    int (*set_features)(struct vhost_dev *vdev, u64 device_features); >>> +    int (*set_status)(struct vhost_dev *vdev, u8 status); >>> +    u8 (*get_status)(struct vhost_dev *vdev); >>> +}; >>> + >>> struct virtio_config_ops >>> I think there's some overlap here and some of the ops tries to do the >>> same thing. >>> >>> I think it differs in (*set_vq_address)() and (*create_vqs)(). >>> [create_vqs() introduced in struct vhost_config_ops provides >>> complimentary functionality to (*find_vqs)() in struct >>> virtio_config_ops. It seemingly encapsulates the functionality of >>> (*set_vq_address)(), (*set_vq_num)(), (*set_vq_cb)(),..]. >>> >>> Back to the difference between (*set_vq_address)() and (*create_vqs)(), >>> set_vq_address() directly provides the virtqueue address to the vdpa >>> device but create_vqs() only provides the parameters of the virtqueue >>> (like the number of virtqueues, number of buffers) but does not directly >>> provide the address. IMO the backend client drivers (like net or vhost) >>> shouldn't/cannot by itself know how to access the vring created on >>> virtio front-end. The vdpa device/vhost device should have logic for >>> that. That will help the client drivers to work with different types of >>> vdpa device/vhost device and can access the vring created by virtio >>> irrespective of whether the vring can be accessed via mmio or kernel >>> space or user space. >>> >>> I think vdpa always works with client drivers in userspace and providing >>> userspace address for vring. >> >> Sorry for being unclear. What I meant is not replacing vDPA with the >> vhost(bus) you proposed but the possibility of replacing virtio-pci-epf >> with vDPA in: > Okay, so the virtio back-end still use vhost and front end should use > vDPA. I see. So the host side PCI driver for EPF should populate > vdpa_config_ops and invoke vdpa_register_device(). Yes. >> My question is basically for the part of virtio_pci_epf_send_command(), >> so it looks to me you have a vendor specific API to replace the >> virtio-pci layout of the BAR: > Even when we use vDPA, we have to use some sort of > virtio_pci_epf_send_command() to communicate with virtio backend right? Right. > > Right, the layout is slightly different from the standard layout. > > This is the layout > struct epf_vhost_reg_queue { > u8 cmd; > u8 cmd_status; > u16 status; > u16 num_buffers; > u16 msix_vector; > u64 queue_addr; What's the meaning of queue_addr here? Does not mean the device expects a contiguous memory for avail/desc/used ring? > } __packed; > > struct epf_vhost_reg { > u64 host_features; > u64 guest_features; > u16 msix_config; > u16 num_queues; > u8 device_status; > u8 config_generation; > u32 isr; > u8 cmd; > u8 cmd_status; > struct epf_vhost_reg_queue vq[MAX_VQS]; > } __packed; >> >> +static int virtio_pci_epf_send_command(struct virtio_pci_device *vp_dev, >> +                       u32 command) >> +{ >> +    struct virtio_pci_epf *pci_epf; >> +    void __iomem *ioaddr; >> +    ktime_t timeout; >> +    bool timedout; >> +    int ret = 0; >> +    u8 status; >> + >> +    pci_epf = to_virtio_pci_epf(vp_dev); >> +    ioaddr = vp_dev->ioaddr; >> + >> +    mutex_lock(&pci_epf->lock); >> +    writeb(command, ioaddr + HOST_CMD); >> +    timeout = ktime_add_ms(ktime_get(), COMMAND_TIMEOUT); >> +    while (1) { >> +        timedout = ktime_after(ktime_get(), timeout); >> +        status = readb(ioaddr + HOST_CMD_STATUS); >> + >> >> Several questions: >> >> - It's not clear to me how the synchronization is done between the RC >> and EP. E.g how and when the value of HOST_CMD_STATUS can be changed. > The HOST_CMD (commands sent to the EP) is serialized by using mutex. > Once the EP reads the command, it resets the value in HOST_CMD. So > HOST_CMD is less likely an issue. Here's my understanding of the protocol: 1) RC write to HOST_CMD 2) RC wait for HOST_CMD_STATUS to be HOST_CMD_STATUS_OKAY It looks to me what EP should do is 1) EP reset HOST_CMD after reading new command And it looks to me EP should also reset HOST_CMD_STATUS here? (I thought there should be patch to handle stuffs like this but I didn't find it in this series) > > A sufficiently large time is given for the EP to complete it's operation > (1 Sec) where the EP provides the status in HOST_CMD_STATUS. After it > expires, HOST_CMD_STATUS_NONE is written to HOST_CMD_STATUS. There could > be case where EP updates HOST_CMD_STATUS after RC writes > HOST_CMD_STATUS_NONE, but by then HOST has already detected this as > failure and error-ed out. > >> If you still want to introduce a new transport, a virtio spec patch >> would be helpful for us to understand the device API. > Okay, that should be on https://github.com/oasis-tcs/virtio-spec.git? Yes. >> - You have you vendor specific layout (according to >> virtio_pci_epb_table()), so I guess you it's better to have a vendor >> specific vDPA driver instead > Okay, with vDPA, we are free to define our own layouts. Right, but vDPA have other requirements. E.g it requires the device have the ability to save/restore the state (e.g the last_avail_idx). So it actually depends on what you want. If you don't care about userspace drivers and want to have a standard transport, you can still go virtio. >> - The advantage of vendor specific vDPA driver is that it can 1) have >> less codes 2) support userspace drivers through vhost-vDPA (instead of >> inventing new APIs since we can't use vfio-pci here). > I see there's an additional level of indirection from virtio to vDPA and > probably no need for spec update but don't exactly see how it'll reduce > code. AFAIK you don't need to implement your own setup_vq and del_vq. > > For 2, Isn't vhost-vdpa supposed to run on virtio backend? Not currently, vDPA is a superset of virtio (e.g it support virtqueue state save/restore). This it should be possible in the future probably. > > From a high level, I think I should be able to use vDPA for > virtio_pci_epf.c. Would you also suggest using vDPA for ntb_virtio.c? > ([RFC PATCH 20/22] NTB: Add a new NTB client driver to implement VIRTIO > functionality). I think it's your call. If you want 1) a well-defined standard virtio transport 2) willing to finalize d and maintain the spec 3) doesn't care about userspace drivers You can go with virtio, otherwise vDPA. Thanks > > Thanks > Kishon >