Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753662Ab0KICPq (ORCPT ); Mon, 8 Nov 2010 21:15:46 -0500 Received: from mx1.redhat.com ([209.132.183.28]:28030 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752107Ab0KICPp (ORCPT ); Mon, 8 Nov 2010 21:15:45 -0500 Date: Mon, 8 Nov 2010 21:15:40 -0500 From: Vivek Goyal To: Shaohua Li Cc: lkml , Jens Axboe , "czoccolo@gmail.com" Subject: Re: [patch 2/3]cfq-iosched: schedule dispatch for noidle queue Message-ID: <20101109021540.GA29210@redhat.com> References: <1289182038.23014.190.camel@sli10-conroe> <20101108142842.GC16767@redhat.com> <1289266267.23014.196.camel@sli10-conroe> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1289266267.23014.196.camel@sli10-conroe> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3588 Lines: 80 On Tue, Nov 09, 2010 at 09:31:07AM +0800, Shaohua Li wrote: > On Mon, 2010-11-08 at 22:28 +0800, Vivek Goyal wrote: > > On Mon, Nov 08, 2010 at 10:07:18AM +0800, Shaohua Li wrote: > > > A queue is idle at cfq_dispatch_requests(), but it gets noidle later. Unless > > > other task explictly does unplug or all requests are drained, we will not > > > deliever requests to the disk even cfq_arm_slice_timer doesn't make the > > > queue idle. For example, cfq_should_idle() returns true because of > > > service_tree->count == 1, and then other queues are added. Note, I didn't > > > see obvious performance impacts so far with the patch, but just thought > > > this could be a problem. > > > > > > Signed-off-by: Shaohua Li > > > > > > --- > > > block/cfq-iosched.c | 23 ++++++++++++++++++++++- > > > 1 file changed, 22 insertions(+), 1 deletion(-) > > > > > > Index: linux/block/cfq-iosched.c > > > =================================================================== > > > --- linux.orig/block/cfq-iosched.c 2010-11-08 08:41:20.000000000 +0800 > > > +++ linux/block/cfq-iosched.c 2010-11-08 08:43:51.000000000 +0800 > > > @@ -3265,6 +3265,10 @@ cfq_should_preempt(struct cfq_data *cfqd > > > if (cfq_class_rt(new_cfqq) && !cfq_class_rt(cfqq)) > > > return true; > > > > > > + /* An idle queue should not be idle now for some reason */ > > > + if (RB_EMPTY_ROOT(&cfqq->sort_list) && !cfq_should_idle(cfqd, cfqq)) > > > + return true; > > > + > > > if (!cfqd->active_cic || !cfq_cfqq_wait_request(cfqq)) > > > return false; > > > > > > @@ -3508,8 +3512,25 @@ static void cfq_completed_request(struct > > > } > > > } > > > > > > - if (!cfqd->rq_in_driver) > > > + if (!cfqd->rq_in_driver) { > > > cfq_schedule_dispatch(cfqd); > > > + return; > > > + } > > > + /* > > > + * A queue is idle at cfq_dispatch_requests(), but it gets noidle > > > + * later. We schedule a dispatch if the queue has no requests, > > > + * otherwise the disk is actually in idle till all requests > > > + * are finished even cfq_arm_slice_timer doesn't make the queue idle > > > + * */ > > > > Why do we have to wait for all requests to finish in device? Will driver > > most likely not ask for next request when 1-2 requests have completed > > and at that time we should expire the queue if queue is no more marked > > as "noidle"? > The issue is a queue is idle just because it's the last queue of the > service tree. Then a new queue is added and the idled queue should not > idle now. we should preempt the idled queue soon. does this make sense > to you? If that's the case then you should just modify should_preempt() so that addition of a new queue could preempt an empty queue which has now become noidle. You have also modified cfq_completed_request() function, which will wake up the worker thread and then try to dispatch a request. IMHO, in practice driver asks for new request almost immediately and you don't gain much by this additional wakeup. So my point being, that we increased the code complexity for no visible performance improvement also increased thread wakeups resulting in more cpu consumption. If there was a visible performance gain in your testing then it would have made sense but you said that you did not notice any improvements. Then why to increase the complexity. Vivek -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/