Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S935894AbZLPXBv (ORCPT ); Wed, 16 Dec 2009 18:01:51 -0500 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S935330AbZLPXB3 (ORCPT ); Wed, 16 Dec 2009 18:01:29 -0500 Received: from mail.vyatta.com ([76.74.103.46]:46098 "EHLO mail.vyatta.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S935408AbZLPXBU (ORCPT ); Wed, 16 Dec 2009 18:01:20 -0500 Date: Wed, 16 Dec 2009 15:00:51 -0800 From: Stephen Hemminger To: David Daney Cc: Chetan Loke , Chris Friesen , netdev@vger.kernel.org, Linux Kernel Mailing List , linux-mips Subject: Re: Irq architecture for multi-core network driver. Message-ID: <20091216150051.63b6e31c@nehalam> In-Reply-To: <4B295F8C.4050905@caviumnetworks.com> References: <4AE0D14B.1070307@caviumnetworks.com> <4AE0D72A.4090607@nortel.com> <4AE0DB98.1000101@caviumnetworks.com> <4B295F8C.4050905@caviumnetworks.com> Organization: Vyatta X-Mailer: Claws Mail 3.7.2 (GTK+ 2.18.3; x86_64-pc-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 2576 Lines: 56 On Wed, 16 Dec 2009 14:30:36 -0800 David Daney wrote: > Chetan Loke wrote: > >>> Does your hardware do flow-based queues? In this model you have > >>> multiple rx queues and the hardware hashes incoming packets to a single > >>> queue based on the addresses, ports, etc. This ensures that all the > >>> packets of a single connection always get processed in the order they > >>> arrived at the net device. > >>> > >> Indeed, this is exactly what we have. > >> > >> > >>> Typically in this model you have as many interrupts as queues > >>> (presumably 16 in your case). Each queue is assigned an interrupt and > >>> that interrupt is affined to a single core. > > > >> Certainly this is one mode of operation that should be supported, but I > >> would also like to be able to go for raw throughput and have as many cores > >> as possible reading from a single queue (like I currently have). > >> > > Well, you could let the NIC firmware(f/w) handle this. The f/w would > > know which interrupt was just injected recently.In other words it > > would have a history of which CPU's would be available. So if some > > previously interrupted CPU isn't making good progress then the > > firmware should route the incoming response packets to a different > > queue. This way some other CPU will pick it up. > > > > > It isn's a NIC. There is no firmware. The system interrupt hardware is > what it is and cannot be changed. > > My current implementation still has a single input queue configured and > I get a maskable interrupt on a single CPU when packets are available. > If the queue depth increases above a given threshold, I optionally send > an IPI to another CPU to enable NAPI polling on that CPU. > > Currently I have a module parameter that controls the maximum number of > CPUs that will have NAPI polling enabled. > > This allows me to get multiple CPUs doing receive processing without > having to hack into the lower levels of the system's interrupt > processing code to try to do interrupt steering. Since all the > interrupt service routine was doing was call netif_rx_schedule(), I can > simply do this via smp_call_function_single(). Better to look into receive packet steering patches that are still under review (rather than reinventing it just for your driver) -- -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/