Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp3256455pxj; Mon, 7 Jun 2021 06:24:22 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz1SzcQpz/AFg0m86s+f4vcCtXSvYuetdAKuDr1p5dNjb1M9Vn+Zw5j32neMWEl9E2BwA94 X-Received: by 2002:a17:906:1682:: with SMTP id s2mr17755942ejd.172.1623072262252; Mon, 07 Jun 2021 06:24:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623072262; cv=none; d=google.com; s=arc-20160816; b=X5ADdAgnIVAFBDElgko4JKX5QXrb42Zm7RiGSfXPmwqKgGlDPVzeSFn0S5Z80td05C AdbiasSPLEiyWE59gVrNLtkWrliQd54e+1hTmPBsgGx1WLuFGmoQ7hst/q+Y4mSYKko+ 8nLFCKHFuBDnVg5s8Jhl81xDgfAYN8uoGgPylMFa6myFkZ3GCRYn+j6E8sVvvlidPsQG HqdJwwC1tHtdain/oauPMnKRTNBdC7FIzU1bOO+MSfWmdr4hcVC7Z48sWPf/y9v9HFFp rEON2JRcibBLTfiZIVBX55q7K1tg/xZ0ToaGS85lF5nhEDA4Cdv8do97od7fOcTg9U97 3aaQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-transfer-encoding :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature; bh=VXxbAyP7IosdKGXXuxy96E3+0YLCtyxIMTIoO1dzAYM=; b=FgnJR9Nn+PJKxqY3h6TX5+GeVOcHJOOKAdYmzirtn5yZqr9IJqCy5/7ivYK4OO1f1Z 3DlDaBPFYb3RxrkzHieH3HK8NXZqxr8xbBZKnryyihry9PcEB04hBmmpzM22u50HKw1M lC50HfqpW0qo04+jHm15cVdYogfAG+pkeuaP7jSgvU0IbdK0eh1K/P4CH8LpeDBtBUYJ /lsBcY5dnszrEswaK9VnrC8jzQiNpskCvLIk+cf97mnCEOPn0RqrngPykzfAnI9r29Ct 9UeCnwmkZswhz+arfwmVO2xIv2OA7tOVwWp11q83m35+MW9QIT7+WxtbQC/jyesAnYTs 1XSA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=Oh6HIdN1; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id dj4si4198827edb.230.2021.06.07.06.23.59; Mon, 07 Jun 2021 06:24:22 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=Oh6HIdN1; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230311AbhFGNWT (ORCPT + 99 others); Mon, 7 Jun 2021 09:22:19 -0400 Received: from mail.kernel.org ([198.145.29.99]:52990 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230239AbhFGNWS (ORCPT ); Mon, 7 Jun 2021 09:22:18 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 1389E61153; Mon, 7 Jun 2021 13:20:27 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1623072027; bh=o3kGqcgRA7pGA6vhHOS2lBK9GrMCiVtqJGaYEdbAMNw=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=Oh6HIdN1VZKdci8pYkMcfgQZwACR2UURtKBf0ZUX8FZWBh+PXevNsj3uBt5VF/gO2 kB+qUNJcA+maXC5IqRB5/sFhktAnIGjPcyGdjvmpRHNADdnW3/5aTvervTSP/vsGMa fyVeKLQsIvG+RPyIDgb80G6kc58kysIB9wV1QFlwfn0Jar4T/kDAUsw3TOFzY8y6hF xtO4oIpRpj/fbaXtXimNp1sWJlURhrqbny9uiO8VASC8TwuTEsaCFQPbak0wyvU7TV n7jDUfkj8KMVmX+wOuboKdDp/wTWX9ZfQint6DyHQbpy+WBp2+RqtBkU1tNarzmt/i DW5DgwdbNd3Gg== Received: by quaco.ghostprotocols.net (Postfix, from userid 1000) id 4E43840B1A; Mon, 7 Jun 2021 10:20:24 -0300 (-03) Date: Mon, 7 Jun 2021 10:20:24 -0300 From: Arnaldo Carvalho de Melo To: Andrii Nakryiko Cc: Andrii Nakryiko , Jiri Olsa , dwarves@vger.kernel.org, bpf , Kernel Team , Linux Kernel Mailing List Subject: Parallelizing vmlinux BTF encoding. was Re: [RFT] Testing 1.22 Message-ID: References: <4615C288-2CFD-483E-AB98-B14A33631E2F@gmail.com> <4901AF88-0354-428B-9305-2EDC6F75C073@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: X-Url: http://acmel.wordpress.com Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Em Fri, Jun 04, 2021 at 07:55:17PM -0700, Andrii Nakryiko escreveu: > On Thu, Jun 3, 2021 at 7:57 AM Arnaldo Carvalho de Melo wrote: > > Em Sat, May 29, 2021 at 05:40:17PM -0700, Andrii Nakryiko escreveu: > > > At some point it probably would make sense to formalize > > > "btf_encoder" as a struct with its own state instead of passing in > > > multiple variables. It would probably also > > Take a look at the tmp.master branch at: > > https://git.kernel.org/pub/scm/devel/pahole/pahole.git/log/?h=tmp.master > Oh wow, that's a lot of commits! :) Great that you decided to do this > refactoring, thanks! > > that btf_elf class isn't used anymore by btf_loader, that uses only > > libbpf's APIs, and now we have a btf_encoder class with all the globals, > > etc, more baby steps are needed to finally ditch btf_elf altogether and > > move on to the parallelization. > So do you plan to try to parallelize as a next step? I'm pretty So, I haven't looked at details but what I thought would be interesting to investigate is to see if we can piggyback DWARF generation with BTF one, i.e. when we generate a .o file with -g we encode the DWARF info, so, right after this, we could call pahole as-is and encode BTF, then, when vmlinux is linked, we would do the dedup. I.e. when generating ../build/v5.13.0-rc4+/kernel/fork.o, that comes with: ⬢[acme@toolbox perf]$ readelf -SW ../build/v5.13.0-rc4+/kernel/fork.o | grep debug [78] .debug_info PROGBITS 0000000000000000 00daec 032968 00 0 0 1 [79] .rela.debug_info RELA 0000000000000000 040458 053b68 18 I 95 78 8 [80] .debug_abbrev PROGBITS 0000000000000000 093fc0 0012e9 00 0 0 1 [81] .debug_loclists PROGBITS 0000000000000000 0952a9 00aa43 00 0 0 1 [82] .rela.debug_loclists RELA 0000000000000000 09fcf0 009d98 18 I 95 81 8 [83] .debug_aranges PROGBITS 0000000000000000 0a9a88 000080 00 0 0 1 [84] .rela.debug_aranges RELA 0000000000000000 0a9b08 0000a8 18 I 95 83 8 [85] .debug_rnglists PROGBITS 0000000000000000 0a9bb0 001509 00 0 0 1 [86] .rela.debug_rnglists RELA 0000000000000000 0ab0c0 001bc0 18 I 95 85 8 [87] .debug_line PROGBITS 0000000000000000 0acc80 0086b7 00 0 0 1 [88] .rela.debug_line RELA 0000000000000000 0b5338 002550 18 I 95 87 8 [89] .debug_str PROGBITS 0000000000000000 0b7888 0177ad 01 MS 0 0 1 [90] .debug_line_str PROGBITS 0000000000000000 0cf035 001308 01 MS 0 0 1 [93] .debug_frame PROGBITS 0000000000000000 0d0370 000e38 00 0 0 8 [94] .rela.debug_frame RELA 0000000000000000 0d11a8 000e70 18 I 95 93 8 ⬢[acme@toolbox perf]$ We would do: ⬢[acme@toolbox perf]$ pahole -J ../build/v5.13.0-rc4+/kernel/fork.o ⬢[acme@toolbox perf]$ Which would get us to have: ⬢[acme@toolbox perf]$ readelf -SW ../build/v5.13.0-rc4+/kernel/fork.o | grep BTF [103] .BTF PROGBITS 0000000000000000 0db658 030550 00 0 0 1 ⬢[acme@toolbox perf] ⬢[acme@toolbox perf]$ pahole -F btf -C hlist_node ../build/v5.13.0-rc4+/kernel/fork.o struct hlist_node { struct hlist_node * next; /* 0 8 */ struct hlist_node * * pprev; /* 8 8 */ /* size: 16, cachelines: 1, members: 2 */ /* last cacheline: 16 bytes */ }; ⬢[acme@toolbox perf]$ So, a 'pahole --dedup_btf vmlinux' would just go on looking at: ⬢[acme@toolbox perf]$ readelf -wi ../build/v5.13.0-rc4+/vmlinux | grep -A10 DW_TAG_compile_unit | grep -w DW_AT_name | grep fork DW_AT_name : (indirect line string, offset: 0x62e7): /var/home/acme/git/linux/kernel/fork.c To go there and go on extracting those ELF sections to combine and dedup. This combine thing could be done even by the linker, I think, when all the DWARF data in the .o file are combined into vmlinux, we could do it for the .BTF sections as well, that way would be even more elegant, I think. Then, the combined vmlinux .BTF section would be read and fed in one go to libbtf's dedup arg. This way the encoding of BTF would be as paralellized as the kernel build process, following the same logic (-j NR_PROCESSORS). wdyt? If this isn't the case, we can process vmlinux as is today and go on creating N threads and feeding each with a DW_TAG_compile_unit "container", i.e. each thread would consume all the tags below each DW_TAG_compile_unit and produce a foo.BTF file that in the end would be combined and deduped by libbpf. Doing it as my first sketch above would take advantage of locality of reference, i.e. the DWARF data would be freshly produced and in the cache hierarchy when we first encode BTF, later, when doing the combine+dedup we wouldn't be touching the more voluminous DWARF data. - Arnaldo > confident about BTF encoding part: dump each CU into its own BTF, use > btf__add_type() to merge multiple BTFs together. Just need to re-map > IDs (libbpf internally has API to visit each field that contains > type_id, it's well-defined enough to expose that as a public API, if > necessary). Then final btf_dedup(). > But the DWARF loading and parsing part is almost a black box to me, so > I'm not sure how much work it would involve. > > I'm doing 'pahole -J vmlinux && btfdiff' after each cset and doing it > > very piecemeal as I'm doing will help bisecting any subtle bug this may > > introduce. > > > allow to parallelize BTF generation, where each CU would proceed in > > > parallel generating local BTF, and then the final pass would merge and > > > dedup BTFs. Currently reading and processing DWARF is the slowest part > > > of the DWARF-to-BTF conversion, parallelization and maybe some other > > > optimization seems like the only way to speed the process up. > > > Acked-by: Andrii Nakryiko > > Thanks!