Received: by 2002:a05:6358:bb9e:b0:b9:5105:a5b4 with SMTP id df30csp5089988rwb; Tue, 6 Sep 2022 18:46:46 -0700 (PDT) X-Google-Smtp-Source: AA6agR5RMFhZgrSFOzCoOilqkcYp3tS9haFXNFu7s3M/LSQbj282RHttfrNS0WZ9lPqw84U2J6K6 X-Received: by 2002:a63:4726:0:b0:430:3886:5a2f with SMTP id u38-20020a634726000000b0043038865a2fmr1341386pga.99.1662515205908; Tue, 06 Sep 2022 18:46:45 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1662515205; cv=none; d=google.com; s=arc-20160816; b=kiwUHCrtiz/Zi3HKwlqSol9X9T0NQXLA22Xpu6NMQMRthQK9DDHFYnji0DD0mQKtdX idZe56nuw9gNdQJWctv2RUpQ7V5VYPnwcb6TYOs5xhIZ02MNioYQqMQoB5Ozsg96GV1x uQjhsJ0Sm6ngHYw1LtiHEqW7GphFPGyuMn6QS6081bbQ60KwHC4VoiWJGbdShKzVqHbk aBGxxa8HG7yuj6LlxxoZfpZGZS6fpyjZz7VODx0ilL1w1IGfiJ+s7OY0sY0SstruJHfX uVA6tP+O1ZejR8SVkubSjODwrFNatKLdULrrwxd2sh4wVV8nZQcUjHrxRKIxE2Z2OXRK imyA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to :mime-version:user-agent:date:message-id:from:references:cc:to :subject; bh=rbKtHD1N71NWJrgWMehgkaHRi0vkPUVr20dZs3HnuvY=; b=0MfJCB9Y1uGlgeVq8OAiqZYKfJIXNUsSZMy5Vk2k0kyqbkMBeloJrj4wXTXCDUlQei 6mcmbt0/7rlnIkV+wgMEavIGwLMqDIrEPcIx8DKX+HbzQOTgbwlnQ1SAborlhMVwgu7C DVOxlFvAz7rBYXMOay4nVTIBO7xwCQI8IGxJ1bZYbAC7pLoedUFCeAfClsRsTTNC3Fzd Ow+JhrWkvLiPnx+8TKoqzfRTi/lbr/a/FkxhFLcesbZ37/fSviqnSqw9KnpCnibylTMz AmwsWwaAcaSriXo8fqr1X4ojXICPibxfgdooTBGNfKSWh1c1YKUyj5kGndSXoWKSeBBT FFLw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id r20-20020aa79634000000b00535fa14ffaasi14295146pfg.116.2022.09.06.18.46.34; Tue, 06 Sep 2022 18:46:45 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229707AbiIGBQb (ORCPT + 99 others); Tue, 6 Sep 2022 21:16:31 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59612 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229449AbiIGBQ3 (ORCPT ); Tue, 6 Sep 2022 21:16:29 -0400 Received: from dggsgout11.his.huawei.com (unknown [45.249.212.51]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D5B9880F5C; Tue, 6 Sep 2022 18:16:27 -0700 (PDT) Received: from mail02.huawei.com (unknown [172.30.67.143]) by dggsgout11.his.huawei.com (SkyGuard) with ESMTP id 4MMknW0pbWzklTn; Wed, 7 Sep 2022 09:14:55 +0800 (CST) Received: from [10.174.176.73] (unknown [10.174.176.73]) by APP2 (Coremail) with SMTP id Syh0CgC3VW_p8BdjSXfqAQ--.36493S3; Wed, 07 Sep 2022 09:16:26 +0800 (CST) Subject: Re: [PATCH -next v10 3/4] block, bfq: refactor the counting of 'num_groups_with_pending_reqs' To: Paolo Valente , Yu Kuai Cc: Jan Kara , cgroups@vger.kernel.org, linux-block , Tejun Heo , Jens Axboe , LKML , yi.zhang@huawei.com, "yukuai (C)" References: <20220610021701.2347602-1-yukuai3@huawei.com> <20220610021701.2347602-4-yukuai3@huawei.com> <27F2DF19-7CC6-42C5-8CEB-43583EB4AE46@linaro.org> <9b2d667f-6636-9347-08a1-8bd0aa2346f2@huaweicloud.com> <2f94f241-445f-1beb-c4a8-73f6efce5af2@huaweicloud.com> <55A07102-BE55-4606-9E32-64E884064FB9@unimore.it> <5cb0e5bc-feec-86d6-6f60-3c28ee625efd@huaweicloud.com> From: Yu Kuai Message-ID: <5e0b44b4-46cc-b3c6-1d93-00a0a683eda8@huaweicloud.com> Date: Wed, 7 Sep 2022 09:16:25 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:60.0) Gecko/20100101 Thunderbird/60.8.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit X-CM-TRANSID: Syh0CgC3VW_p8BdjSXfqAQ--.36493S3 X-Coremail-Antispam: 1UD129KBjvJXoWfGF4rZF4xur1UXr43GFy8Xwb_yoWDtryrpr W5Ga1UAr4UXr1UJr1Utr1UXr10q34fJryUXr1DJr1UAr1qyFn7JF17tr18ury8Xr18Jr12 qr1UXr9rXr1UJFDanT9S1TB71UUUUUUqnTZGkaVYY2UrUUUUjbIjqfuFe4nvWSU5nxnvy2 9KBjDU0xBIdaVrnRJUUU9014x267AKxVW8JVW5JwAFc2x0x2IEx4CE42xK8VAvwI8IcIk0 rVWrJVCq3wAFIxvE14AKwVWUJVWUGwA2ocxC64kIII0Yj41l84x0c7CEw4AK67xGY2AK02 1l84ACjcxK6xIIjxv20xvE14v26F1j6w1UM28EF7xvwVC0I7IYx2IY6xkF7I0E14v26F4j 6r4UJwA2z4x0Y4vEx4A2jsIE14v26rxl6s0DM28EF7xvwVC2z280aVCY1x0267AKxVW0oV Cq3wAS0I0E0xvYzxvE52x082IY62kv0487Mc02F40EFcxC0VAKzVAqx4xG6I80ewAv7VC0 I7IYx2IY67AKxVWUGVWUXwAv7VC2z280aVAFwI0_Jr0_Gr1lOx8S6xCaFVCjc4AY6r1j6r 4UM4x0Y48IcVAKI48JM4x0x7Aq67IIx4CEVc8vx2IErcIFxwACI402YVCY1x02628vn2kI c2xKxwCYjI0SjxkI62AI1cAE67vIY487MxAIw28IcxkI7VAKI48JMxC20s026xCaFVCjc4 AY6r1j6r4UMI8I3I0E5I8CrVAFwI0_Jr0_Jr4lx2IqxVCjr7xvwVAFwI0_JrI_JrWlx4CE 17CEb7AF67AKxVWUtVW8ZwCIc40Y0x0EwIxGrwCI42IY6xIIjxv20xvE14v26r1j6r1xMI IF0xvE2Ix0cI8IcVCY1x0267AKxVW8JVWxJwCI42IY6xAIw20EY4v20xvaj40_Zr0_Wr1U MIIF0xvEx4A2jsIE14v26r1j6r4UMIIF0xvEx4A2jsIEc7CjxVAFwI0_Gr0_Gr1UYxBIda VFxhVjvjDU0xZFpf9x0JUZa9-UUUUU= X-CM-SenderInfo: 51xn3trlr6x35dzhxuhorxvhhfrp/ X-CFilter-Loop: Reflected X-Spam-Status: No, score=-2.7 required=5.0 tests=BAYES_00,KHOP_HELO_FCRDNS, MAY_BE_FORGED,NICE_REPLY_A,SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, Paolo! 在 2022/09/06 17:37, Paolo Valente 写道: > > >> Il giorno 26 ago 2022, alle ore 04:34, Yu Kuai ha scritto: >> >> Hi, Paolo! >> >> 在 2022/08/25 22:59, Paolo Valente 写道: >>>> Il giorno 11 ago 2022, alle ore 03:19, Yu Kuai > ha scritto: >>>> >>>> Hi, Paolo >>>> >>>> 在 2022/08/10 18:49, Paolo Valente 写道: >>>>>> Il giorno 27 lug 2022, alle ore 14:11, Yu Kuai > ha scritto: >>>>>> >>>>>> Hi, Paolo >>>>>> >>>>> hi >>>>>> Are you still interested in this patchset? >>>>>> >>>>> Yes. Sorry for replying very late again. >>>>> Probably the last fix that you suggest is enough, but I'm a little bit >>>>> concerned that it may be a little hasty. In fact, before this fix, we >>>>> exchanged several messages, and I didn't seem to be very good at >>>>> convincing you about the need to keep into account also in-service >>>>> I/O. So, my question is: are you sure that now you have a >>>> >>>> I'm confused here, I'm pretty aware that in-service I/O(as said pending >>>> requests is the patchset) should be counted, as you suggested in v7, are >>>> you still thinking that the way in this patchset is problematic? >>>> >>>> I'll try to explain again that how to track is bfqq has pending pending >>>> requests, please let me know if you still think there are some problems: >>>> >>>> patch 1 support to track if bfqq has pending requests, it's >>>> done by setting the flag 'entity->in_groups_with_pending_reqs' when the >>>> first request is inserted to bfqq, and it's cleared when the last >>>> request is completed. specifically the flag is set in >>>> bfq_add_bfqq_busy() when 'bfqq->dispatched' if false, and it's cleared >>>> both in bfq_completed_request() and bfq_del_bfqq_busy() when >>>> 'bfqq->diapatched' is false. >>>> >>> This general description seems correct to me. Have you already sent a new version of your patchset? >> >> It's glad that we finially on the same page here. >> > > Yep. Sorry for my chronicle delay. Better late than never ???? > >> Please take a look at patch 1, which already impelement the above >> descriptions, it seems to me there is no need to send a new version >> for now. If you think there are still some other problems, please let >> me know. >> > > Patch 1 seems ok to me. I seem to have only one pending comment on this patch (3/4) instead. Let me paste previous stuff here for your convenience: That sounds good. > >>> >>> - /* >>> - * Next function is invoked last, because it causes bfqq to be >>> - * freed if the following holds: bfqq is not in service and >>> - * has no dispatched request. DO NOT use bfqq after the next >>> - * function invocation. >>> - */ >> I would really love it if you leave this comment. I added it after >> suffering a lot for a nasty UAF. Of course the first sentence may >> need to be adjusted if the code that precedes it is to be removed. >> Same as above, if this patch is applied, this function will be gone. > > yes, but this comment now must be moved forward. > > Looking forward for a new complete version, for a new review. I'll do > my best to reply quicker. I'll send a new version soon, perhaps I'll also add the following cleanups to this patchset. Thanks, Kuai > > Thanks, > Paolo > > > > > >> Thanks, >> Kuai >>> Thanks, >>> Paolo >>>> Thanks, >>>> Kuai >>>>> clear/complete understanding of this non-trivial matter? >>>>> Consequently, are we sure that this last fix is most certainly all we >>>>> need? Of course, I will check on my own, but if you reassure me on >>>>> this point, I will feel more confident. >>>>> Thanks, >>>>> Paolo >>>>>> 在 2022/07/20 19:38, Yu Kuai 写道: >>>>>>> Hi >>>>>>> >>>>>>> 在 2022/07/20 19:24, Paolo VALENTE 写道: >>>>>>>> >>>>>>>> >>>>>>>>> Il giorno 12 lug 2022, alle ore 15:30, Yu Kuai > ha scritto: >>>>>>>>> >>>>>>>>> Hi! >>>>>>>>> >>>>>>>>> I'm copying my reply with new mail address, because Paolo seems >>>>>>>>> didn't receive my reply. >>>>>>>>> >>>>>>>>> 在 2022/06/23 23:32, Paolo Valente 写道: >>>>>>>>>> Sorry for the delay. >>>>>>>>>>> Il giorno 10 giu 2022, alle ore 04:17, Yu Kuai > ha scritto: >>>>>>>>>>> >>>>>>>>>>> Currently, bfq can't handle sync io concurrently as long as they >>>>>>>>>>> are not issued from root group. This is because >>>>>>>>>>> 'bfqd->num_groups_with_pending_reqs > 0' is always true in >>>>>>>>>>> bfq_asymmetric_scenario(). >>>>>>>>>>> >>>>>>>>>>> The way that bfqg is counted into 'num_groups_with_pending_reqs': >>>>>>>>>>> >>>>>>>>>>> Before this patch: >>>>>>>>>>> 1) root group will never be counted. >>>>>>>>>>> 2) Count if bfqg or it's child bfqgs have pending requests. >>>>>>>>>>> 3) Don't count if bfqg and it's child bfqgs complete all the requests. >>>>>>>>>>> >>>>>>>>>>> After this patch: >>>>>>>>>>> 1) root group is counted. >>>>>>>>>>> 2) Count if bfqg have pending requests. >>>>>>>>>>> 3) Don't count if bfqg complete all the requests. >>>>>>>>>>> >>>>>>>>>>> With this change, the occasion that only one group is activated can be >>>>>>>>>>> detected, and next patch will support concurrent sync io in the >>>>>>>>>>> occasion. >>>>>>>>>>> >>>>>>>>>>> Signed-off-by: Yu Kuai > >>>>>>>>>>> Reviewed-by: Jan Kara > >>>>>>>>>>> --- >>>>>>>>>>> block/bfq-iosched.c | 42 ------------------------------------------ >>>>>>>>>>> block/bfq-iosched.h | 18 +++++++++--------- >>>>>>>>>>> block/bfq-wf2q.c | 19 ++++--------------- >>>>>>>>>>> 3 files changed, 13 insertions(+), 66 deletions(-) >>>>>>>>>>> >>>>>>>>>>> diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c >>>>>>>>>>> index 0ec21018daba..03b04892440c 100644 >>>>>>>>>>> --- a/block/bfq-iosched.c >>>>>>>>>>> +++ b/block/bfq-iosched.c >>>>>>>>>>> @@ -970,48 +970,6 @@ void __bfq_weights_tree_remove(struct bfq_data *bfqd, >>>>>>>>>>> void bfq_weights_tree_remove(struct bfq_data *bfqd, >>>>>>>>>>> struct bfq_queue *bfqq) >>>>>>>>>>> { >>>>>>>>>>> -struct bfq_entity *entity = bfqq->entity.parent; >>>>>>>>>>> - >>>>>>>>>>> -for_each_entity(entity) { >>>>>>>>>>> -struct bfq_sched_data *sd = entity->my_sched_data; >>>>>>>>>>> - >>>>>>>>>>> -if (sd->next_in_service || sd->in_service_entity) { >>>>>>>>>>> -/* >>>>>>>>>>> -* entity is still active, because either >>>>>>>>>>> -* next_in_service or in_service_entity is not >>>>>>>>>>> -* NULL (see the comments on the definition of >>>>>>>>>>> -* next_in_service for details on why >>>>>>>>>>> -* in_service_entity must be checked too). >>>>>>>>>>> -* >>>>>>>>>>> -* As a consequence, its parent entities are >>>>>>>>>>> -* active as well, and thus this loop must >>>>>>>>>>> -* stop here. >>>>>>>>>>> -*/ >>>>>>>>>>> -break; >>>>>>>>>>> -} >>>>>>>>>>> - >>>>>>>>>>> -/* >>>>>>>>>>> -* The decrement of num_groups_with_pending_reqs is >>>>>>>>>>> -* not performed immediately upon the deactivation of >>>>>>>>>>> -* entity, but it is delayed to when it also happens >>>>>>>>>>> -* that the first leaf descendant bfqq of entity gets >>>>>>>>>>> -* all its pending requests completed. The following >>>>>>>>>>> -* instructions perform this delayed decrement, if >>>>>>>>>>> -* needed. See the comments on >>>>>>>>>>> -* num_groups_with_pending_reqs for details. >>>>>>>>>>> -*/ >>>>>>>>>>> -if (entity->in_groups_with_pending_reqs) { >>>>>>>>>>> -entity->in_groups_with_pending_reqs = false; >>>>>>>>>>> -bfqd->num_groups_with_pending_reqs--; >>>>>>>>>>> -} >>>>>>>>>>> -} >>>>>>>>>> With this part removed, I'm missing how you handle the following >>>>>>>>>> sequence of events: >>>>>>>>>> 1. a queue Q becomes non busy but still has dispatched requests, so >>>>>>>>>> it must not be removed from the counter of queues with pending reqs >>>>>>>>>> yet >>>>>>>>>> 2. the last request of Q is completed with Q being still idle (non >>>>>>>>>> busy). At this point Q must be removed from the counter. It seems to >>>>>>>>>> me that this case is not handled any longer >>>>>>>>> Hi, Paolo >>>>>>>>> >>>>>>>>> 1) At first, patch 1 support to track if bfqq has pending requests, it's >>>>>>>>> done by setting the flag 'entity->in_groups_with_pending_reqs' when the >>>>>>>>> first request is inserted to bfqq, and it's cleared when the last >>>>>>>>> request is completed(based on weights_tree insertion and removal). >>>>>>>>> >>>>>>>> >>>>>>>> In patch 1 I don't see the flag cleared for the request-completion event :( >>>>>>>> >>>>>>>> The piece of code involved is this: >>>>>>>> >>>>>>>> static void bfq_completed_request(struct bfq_queue *bfqq, struct bfq_data *bfqd) >>>>>>>> { >>>>>>>> u64 now_ns; >>>>>>>> u32 delta_us; >>>>>>>> >>>>>>>> bfq_update_hw_tag(bfqd); >>>>>>>> >>>>>>>> bfqd->rq_in_driver[bfqq->actuator_idx]--; >>>>>>>> bfqd->tot_rq_in_driver--; >>>>>>>> bfqq->dispatched--; >>>>>>>> >>>>>>>> if (!bfqq->dispatched && !bfq_bfqq_busy(bfqq)) { >>>>>>>> /* >>>>>>>> * Set budget_timeout (which we overload to store the >>>>>>>> * time at which the queue remains with no backlog and >>>>>>>> * no outstanding request; used by the weight-raising >>>>>>>> * mechanism). >>>>>>>> */ >>>>>>>> bfqq->budget_timeout = jiffies; >>>>>>>> >>>>>>>> bfq_weights_tree_remove(bfqd, bfqq); >>>>>>>> } >>>>>>>> ... >>>>>>>> >>>>>>>> Am I missing something? >>>>>>> >>>>>>> I add a new api bfq_del_bfqq_in_groups_with_pending_reqs() in patch 1 >>>>>>> to clear the flag, and it's called both from bfq_del_bfqq_busy() and >>>>>>> bfq_completed_request(). I think you may miss the later: >>>>>>> >>>>>>> diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c >>>>>>> index 0d46cb728bbf..0ec21018daba 100644 >>>>>>> --- a/block/bfq-iosched.c >>>>>>> +++ b/block/bfq-iosched.c >>>>>>> @@ -6263,6 +6263,7 @@ static void bfq_completed_request(struct bfq_queue *bfqq, struct bfq_data *bfqd) >>>>>>> */ >>>>>>> bfqq->budget_timeout = jiffies; >>>>>>> >>>>>>> + bfq_del_bfqq_in_groups_with_pending_reqs(bfqq); >>>>>>> bfq_weights_tree_remove(bfqd, bfqq); >>>>>>> } >>>>>>> >>>>>>> Thanks, >>>>>>> Kuai >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Paolo >>>>>> >>>>> . >> > > . >