Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932925Ab3JNU3G (ORCPT ); Mon, 14 Oct 2013 16:29:06 -0400 Received: from charlotte.tuxdriver.com ([70.61.120.58]:57820 "EHLO smtp.tuxdriver.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S932266Ab3JNU3D (ORCPT ); Mon, 14 Oct 2013 16:29:03 -0400 Date: Mon, 14 Oct 2013 16:28:54 -0400 From: Neil Horman To: Ingo Molnar Cc: linux-kernel@vger.kernel.org, sebastien.dugue@bull.net, Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , x86@kernel.org Subject: Re: [PATCH] x86: Run checksumming in parallel accross multiple alu's Message-ID: <20131014202854.GH26880@hmsreliant.think-freely.org> References: <1381510298-20572-1-git-send-email-nhorman@tuxdriver.com> <20131012172124.GA18241@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <20131012172124.GA18241@gmail.com> User-Agent: Mutt/1.5.21 (2010-09-15) X-Spam-Score: -2.9 (--) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1802 Lines: 43 On Sat, Oct 12, 2013 at 07:21:24PM +0200, Ingo Molnar wrote: > > * Neil Horman wrote: > > > S?bastien Dugu? reported to me that devices implementing ipoib (which > > don't have checksum offload hardware were spending a significant amount > > of time computing checksums. We found that by splitting the checksum > > computation into two separate streams, each skipping successive elements > > of the buffer being summed, we could parallelize the checksum operation > > accros multiple alus. Since neither chain is dependent on the result of > > the other, we get a speedup in execution (on hardware that has multiple > > alu's available, which is almost ubiquitous on x86), and only a > > negligible decrease on hardware that has only a single alu (an extra > > addition is introduced). Since addition in commutative, the result is > > the same, only faster > > This patch should really come with measurement numbers: what performance > increase (and drop) did you get on what CPUs. > > Thanks, > > Ingo > So, early testing results today. I wrote a test module that, allocated a 4k buffer, initalized it with random data, and called csum_partial on it 100000 times, recording the time at the start and end of that loop. Results on a 2.4 GHz Intel Xeon processor: Without patch: Average execute time for csum_partial was 808 ns With patch: Average execute time for csum_partial was 438 ns I'm looking into hpa's suggestion to use alternate instructions where available right now. I'll have more soon Neil -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/