Received: by 2002:a25:8b12:0:0:0:0:0 with SMTP id i18csp5202529ybl; Tue, 27 Aug 2019 00:41:16 -0700 (PDT) X-Google-Smtp-Source: APXvYqw+PfoDt23sKvnRA1iGmD+UL1doNpMnI7ZlOjNTX8qLIJAnX4AWtBn90wYeJx6Cw5Ovu9Sy X-Received: by 2002:a63:1b66:: with SMTP id b38mr20133253pgm.54.1566891676256; Tue, 27 Aug 2019 00:41:16 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1566891676; cv=none; d=google.com; s=arc-20160816; b=Q//Qt7CYQ2WUyBchsJ+Xgr75qwnSRh2NGNCu5SgQFiFTCa23Slx9K8uXKyjDvwTdTf J2QUb/YsshCOaKaaugSWylGtIrboCJKulh4zkKPGNQBkdSjffFGbba7c4mJussrI8F25 jJ0R65h1jgV0alzo/Xn/csJHkUqV3JmBMSY7KQsP5bvkxNpFJpbwC2FerqGNAw6loWra avI7bFKSrjpf729ylkuje1ka/b1KTyXld+0eSIZHwRKJszahVOVkvTqKrtnV3/k5iI7i OSltowqi+0XQxuVTjcG7bp9Jes8t7xb74scM5vWcquQkhaU3gyGGREKspLnr/5Ran7qP Ee4Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=Q7/Ysi6i3hSX2h3ET8u99u1P+8TfVQZkXvhJxsOQMLY=; b=0Kah7dO0jCNCN4kbBN9m8UBUP/64BRcpS7bM79jrzHfl/0hoi+XReuzKz3Eg+wCq3j Q/JNlCEQbvni0ooBVVcIhv+ab+nxu+4cq1tFX5YLl1Zy1eDHf1/7+MOHXaTAbx/2w/pm AUwXYB18J6uXpHVOA22ha3ODj3KZYcm99GWpYG2DcW8UERa3j9fhfPmDAeXsz0BdoPMq yCOYIVyelaaJr0BgKAM/ohHFNrc9p/0s9A5y2joSbZBsLZ35O/XpZnxlNuvxhZ12aAwV nWeGB6lX5wmRAnyculbrGCxQfyPgZRCUI4YnISvfNfCREPWl7a0xL+WGJWhkF4/CekUs 7RJA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b=gKS6ZjD+; spf=pass (google.com: best guess record for domain of linux-wireless-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-wireless-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id m11si11840771pls.148.2019.08.27.00.41.01; Tue, 27 Aug 2019 00:41:16 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-wireless-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b=gKS6ZjD+; spf=pass (google.com: best guess record for domain of linux-wireless-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-wireless-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1725985AbfH0HlA (ORCPT + 99 others); Tue, 27 Aug 2019 03:41:00 -0400 Received: from mail-qt1-f194.google.com ([209.85.160.194]:35095 "EHLO mail-qt1-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725825AbfH0HlA (ORCPT ); Tue, 27 Aug 2019 03:41:00 -0400 Received: by mail-qt1-f194.google.com with SMTP id u34so20460833qte.2 for ; Tue, 27 Aug 2019 00:40:59 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=Q7/Ysi6i3hSX2h3ET8u99u1P+8TfVQZkXvhJxsOQMLY=; b=gKS6ZjD++I/lhiVWy1yDe4pJ2htQR4MCE5twbK3y9MY+Yzy0RwgEJzLocX61L0AEqr EnuBuwMG14ALQIPdi+dBS3zbTuv4/w3aXtshDG4dgaxfu6OZyc/HsUzRt3i5QKO30HxG Q4NjRApsdWsU/1DPoRsgSu2reyTXdEw0JgWPI= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=Q7/Ysi6i3hSX2h3ET8u99u1P+8TfVQZkXvhJxsOQMLY=; b=cKEs0RspYIE4U/YJH5j3A/buwCC1G/iquazuQs1jG6BVNVqP15CcM9LOT1dk83FySu Yzj7nMoT65r4XewmOgnSdCtjUnB+neOVJIDnR9IbGJPjEuxuF4lMhwqvmsdS/+HU8B6I sxgn6eMdKKrjmcbrsUVZYecEA8ntw7qn6w5XM4P1KkzJmnUKqUgBPwOZihFCEfsF+f8j bSWS1s3HlrEMcceSsCdBw/Pkspu4q+OubB7FgEejjkNifpOk/IiX/taeuM1qYZf3BzXx EXOfSqAs+Z361QuBWkU/iSzu7XLKq56tzlZgvbnDDMCxEwIXJts04kZ7aIM0Vi/hPod0 epsQ== X-Gm-Message-State: APjAAAX+1uI8SG9RAjIm/DQzrLx5ecY/ohbKpAhbrJGPo/pFRLNl3/5D QB20hvg8ghxfZ7d6M4GjsgXMrc6fMizjhlmenFlwKA== X-Received: by 2002:ac8:f43:: with SMTP id l3mr21376787qtk.278.1566891658662; Tue, 27 Aug 2019 00:40:58 -0700 (PDT) MIME-Version: 1.0 References: <1566302108-18219-1-git-send-email-wgong@codeaurora.org> <1566302108-18219-2-git-send-email-wgong@codeaurora.org> In-Reply-To: <1566302108-18219-2-git-send-email-wgong@codeaurora.org> From: Nicolas Boichat Date: Tue, 27 Aug 2019 15:40:47 +0800 Message-ID: Subject: Re: [PATCH 1/7] ath10k: enable RX bundle receive for sdio To: Wen Gong Cc: ath10k@lists.infradead.org, "open list:NETWORKING DRIVERS (WIRELESS)" Content-Type: text/plain; charset="UTF-8" Sender: linux-wireless-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-wireless@vger.kernel.org On Tue, Aug 20, 2019 at 7:55 PM Wen Gong wrote: > > From: Alagu Sankar > > The existing implementation of initiating multiple sdio transfers for > receive bundling is slowing down the receive speed. Combining the > transfers using a bundle method would be ideal. > > The transmission utilization ratio for sdio bus for small packet is > slow, because the space and time cost for sdio bus is same for large > length packet and small length packet. So the speed of data for large > length packet is higher than small length. > > Test result of different length of data: > data packet(byte) cost time(us) calculated rate(Mbps) > 256 28 73 > 512 33 124 > 1024 35 234 > 1792 45 318 > 14336 168 682 > 28672 333 688 > 57344 660 695 > > Tested with QCA6174 SDIO with firmware > WLAN.RMH.4.4.1-00007-QCARMSWP-1. > > Signed-off-by: Alagu Sankar > Signed-off-by: Wen Gong > --- > drivers/net/wireless/ath/ath10k/sdio.c | 101 ++++++++++++++++++++++++--------- > drivers/net/wireless/ath/ath10k/sdio.h | 7 ++- > 2 files changed, 79 insertions(+), 29 deletions(-) > > diff --git a/drivers/net/wireless/ath/ath10k/sdio.c b/drivers/net/wireless/ath/ath10k/sdio.c > index 8ed4fbd..d9395f0 100644 > --- a/drivers/net/wireless/ath/ath10k/sdio.c > +++ b/drivers/net/wireless/ath/ath10k/sdio.c > @@ -24,6 +24,9 @@ > #include "trace.h" > #include "sdio.h" > > +#define ATH10K_SDIO_DMA_BUF_SIZE (32 * 1024) > +#define ATH10K_SDIO_VSG_BUF_SIZE (32 * 1024) > + > /* inlined helper functions */ > > static inline int ath10k_sdio_calc_txrx_padded_len(struct ath10k_sdio *ar_sdio, > @@ -489,11 +492,11 @@ static int ath10k_sdio_mbox_rx_process_packets(struct ath10k *ar, > return ret; > } > > -static int ath10k_sdio_mbox_alloc_pkt_bundle(struct ath10k *ar, > - struct ath10k_sdio_rx_data *rx_pkts, > - struct ath10k_htc_hdr *htc_hdr, > - size_t full_len, size_t act_len, > - size_t *bndl_cnt) > +static int ath10k_sdio_mbox_alloc_bundle(struct ath10k *ar, > + struct ath10k_sdio_rx_data *rx_pkts, > + struct ath10k_htc_hdr *htc_hdr, > + size_t full_len, size_t act_len, > + size_t *bndl_cnt) > { > int ret, i; > > @@ -534,6 +537,7 @@ static int ath10k_sdio_mbox_rx_alloc(struct ath10k *ar, > size_t full_len, act_len; > bool last_in_bundle; > int ret, i; > + int pkt_cnt = 0; > > if (n_lookaheads > ATH10K_SDIO_MAX_RX_MSGS) { > ath10k_warn(ar, > @@ -577,20 +581,22 @@ static int ath10k_sdio_mbox_rx_alloc(struct ath10k *ar, > */ > size_t bndl_cnt; > > - ret = ath10k_sdio_mbox_alloc_pkt_bundle(ar, > - &ar_sdio->rx_pkts[i], > - htc_hdr, > - full_len, > - act_len, > - &bndl_cnt); > + struct ath10k_sdio_rx_data *rx_pkts = > + &ar_sdio->rx_pkts[pkt_cnt]; > + > + ret = ath10k_sdio_mbox_alloc_bundle(ar, > + rx_pkts, > + htc_hdr, > + full_len, > + act_len, > + &bndl_cnt); > > if (ret) { > ath10k_warn(ar, "alloc_bundle error %d\n", ret); > goto err; > } > > - n_lookaheads += bndl_cnt; > - i += bndl_cnt; > + pkt_cnt += bndl_cnt; > /*Next buffer will be the last in the bundle */ > last_in_bundle = true; > } > @@ -602,7 +608,7 @@ static int ath10k_sdio_mbox_rx_alloc(struct ath10k *ar, > if (htc_hdr->flags & ATH10K_HTC_FLAGS_RECV_1MORE_BLOCK) > full_len += ATH10K_HIF_MBOX_BLOCK_SIZE; > > - ret = ath10k_sdio_mbox_alloc_rx_pkt(&ar_sdio->rx_pkts[i], > + ret = ath10k_sdio_mbox_alloc_rx_pkt(&ar_sdio->rx_pkts[pkt_cnt], > act_len, > full_len, > last_in_bundle, > @@ -611,9 +617,10 @@ static int ath10k_sdio_mbox_rx_alloc(struct ath10k *ar, > ath10k_warn(ar, "alloc_rx_pkt error %d\n", ret); > goto err; > } > + pkt_cnt++; > } > > - ar_sdio->n_rx_pkts = i; > + ar_sdio->n_rx_pkts = pkt_cnt; > > return 0; > > @@ -627,41 +634,72 @@ static int ath10k_sdio_mbox_rx_alloc(struct ath10k *ar, > return ret; > } > > -static int ath10k_sdio_mbox_rx_packet(struct ath10k *ar, > - struct ath10k_sdio_rx_data *pkt) > +static int ath10k_sdio_mbox_rx_fetch(struct ath10k *ar) > { > struct ath10k_sdio *ar_sdio = ath10k_sdio_priv(ar); > - struct sk_buff *skb = pkt->skb; > + struct ath10k_sdio_rx_data *pkt = &ar_sdio->rx_pkts[0]; > + struct sk_buff *skb; > int ret; > > + skb = pkt->skb; The code was a bit more compat before, why not keep this? struct sk_buff *skb = pkt->skb; > ret = ath10k_sdio_readsb(ar, ar_sdio->mbox_info.htc_addr, > skb->data, pkt->alloc_len); > - pkt->status = ret; > - if (!ret) > + > + if (ret) { > + ar_sdio->n_rx_pkts = 0; > + ath10k_sdio_mbox_free_rx_pkt(pkt); > + } else { > + pkt->status = ret; > skb_put(skb, pkt->act_len); > + } > > return ret; > } > > -static int ath10k_sdio_mbox_rx_fetch(struct ath10k *ar) > +static int ath10k_sdio_mbox_rx_fetch_bundle(struct ath10k *ar) > { > struct ath10k_sdio *ar_sdio = ath10k_sdio_priv(ar); > + struct ath10k_sdio_rx_data *pkt; > int ret, i; > + u32 pkt_offset, virt_pkt_len; > > + virt_pkt_len = 0; > for (i = 0; i < ar_sdio->n_rx_pkts; i++) { > - ret = ath10k_sdio_mbox_rx_packet(ar, > - &ar_sdio->rx_pkts[i]); > - if (ret) > + virt_pkt_len += ar_sdio->rx_pkts[i].alloc_len; > + } > + > + if (virt_pkt_len < ATH10K_SDIO_DMA_BUF_SIZE) { > + ret = ath10k_sdio_readsb(ar, ar_sdio->mbox_info.htc_addr, > + ar_sdio->vsg_buffer, virt_pkt_len); > + if (ret) { > + i = 0; > goto err; > + } > + } else { > + ath10k_err(ar, "size exceeding limit %d\n", virt_pkt_len); > + } > + > + pkt_offset = 0; > + for (i = 0; i < ar_sdio->n_rx_pkts; i++) { > + struct sk_buff *skb = ar_sdio->rx_pkts[i].skb; > + > + pkt = &ar_sdio->rx_pkts[i]; > + memcpy(skb->data, ar_sdio->vsg_buffer + pkt_offset, > + pkt->alloc_len); Why do you copy alloc_len if you only plan to use act_len? Actually, just use skb_put_data. Also, do you have the same issue as https://patchwork.kernel.org/patch/11116215/ w.r.t. act_len being incorrect? > + pkt->status = 0; > + skb_put(skb, pkt->act_len); > + pkt_offset += pkt->alloc_len; > } > > return 0; > > err: > /* Free all packets that was not successfully fetched. */ > - for (; i < ar_sdio->n_rx_pkts; i++) > + for (i = 0; i < ar_sdio->n_rx_pkts; i++) > ath10k_sdio_mbox_free_rx_pkt(&ar_sdio->rx_pkts[i]); > > + ar_sdio->n_rx_pkts = 0; > + > return ret; > } > > @@ -704,7 +742,10 @@ static int ath10k_sdio_mbox_rxmsg_pending_handler(struct ath10k *ar, > */ > *done = false; > > - ret = ath10k_sdio_mbox_rx_fetch(ar); > + if (ar_sdio->n_rx_pkts > 1) > + ret = ath10k_sdio_mbox_rx_fetch_bundle(ar); > + else > + ret = ath10k_sdio_mbox_rx_fetch(ar); > > /* Process fetched packets. This will potentially update > * n_lookaheads depending on if the packets contain lookahead > @@ -1112,7 +1153,7 @@ static int ath10k_sdio_bmi_get_rx_lookahead(struct ath10k *ar) > MBOX_HOST_INT_STATUS_ADDRESS, > &rx_word); > if (ret) { > - ath10k_warn(ar, "unable to read RX_LOOKAHEAD_VALID: %d\n", ret); > + ath10k_warn(ar, "unable to read rx_lookahd: %d\n", ret); > return ret; > } > > @@ -2007,6 +2048,12 @@ static int ath10k_sdio_probe(struct sdio_func *func, > goto err_core_destroy; > } > > + ar_sdio->vsg_buffer = devm_kmalloc(ar->dev, ATH10K_SDIO_VSG_BUF_SIZE, GFP_KERNEL); > + if (!ar_sdio->vsg_buffer) { > + ret = -ENOMEM; > + goto err_core_destroy; > + } > + > ar_sdio->irq_data.irq_en_reg = > devm_kzalloc(ar->dev, sizeof(struct ath10k_sdio_irq_enable_regs), > GFP_KERNEL); > diff --git a/drivers/net/wireless/ath/ath10k/sdio.h b/drivers/net/wireless/ath/ath10k/sdio.h > index b8c7ac0..4896eca 100644 > --- a/drivers/net/wireless/ath/ath10k/sdio.h > +++ b/drivers/net/wireless/ath/ath10k/sdio.h > @@ -138,8 +138,8 @@ struct ath10k_sdio_irq_proc_regs { > u8 rx_lookahead_valid; > u8 host_int_status2; > u8 gmbox_rx_avail; > - __le32 rx_lookahead[2]; > - __le32 rx_gmbox_lookahead_alias[2]; > + __le32 rx_lookahead[2 * ATH10K_HIF_MBOX_NUM_MAX]; > + __le32 int_status_enable; > }; > > struct ath10k_sdio_irq_enable_regs { > @@ -196,6 +196,9 @@ struct ath10k_sdio { > struct ath10k *ar; > struct ath10k_sdio_irq_data irq_data; > > + /* temporary buffer for sdio read */ > + u8 *vsg_buffer; > + > /* temporary buffer for BMI requests */ > u8 *bmi_buf; > > -- > 1.9.1 >