Received: by 2002:ac0:a5a6:0:0:0:0:0 with SMTP id m35-v6csp6126522imm; Mon, 27 Aug 2018 10:07:25 -0700 (PDT) X-Google-Smtp-Source: ANB0VdbUS2HitMGEc9gI30HXXcwMHvpR8tRbriUU8mse52wAGb/oLy6KX8D+DW8lERLbXBQrd3HQ X-Received: by 2002:a63:2c8e:: with SMTP id s136-v6mr13489601pgs.390.1535389645740; Mon, 27 Aug 2018 10:07:25 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1535389645; cv=none; d=google.com; s=arc-20160816; b=RiFLCYFGZq5vIow852zOijEdQFwdwMSgYfF6jgAH6a0nForB1F/g1HFZlKdliK89x8 uHFcVASgEM+Yq+nTpaMjEU7nKPZ6W15CcKhhE8JfQX5PJ+5++3BwYAGbk4e/sVw/xNDr s80f+tkgbQr9c1gUF1qfU7C9cC5DgcLu5p9c+pzuGLh/lXOmeKlh0VSQnqkN/RqO/aIy iHM0dufEianU7D0lknhuhYX1FGr3zPgsLz3ltrdal5GIU4EdbEUTt9u5bSWI8ZW1W+wC NMCVaBvVKRGOWWM1bCsTskuUaUJeA0pn8vGip54MaMFuef7oUMmZ8dx2c03MoVROadk1 YbSg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:arc-authentication-results; bh=dVZGiEDBGJO0pIe8RJPZppqI2MugvWcBVlIK+kYNEzY=; b=dtwe4UkJ/sMOquWrugsRLEOltXohKbHQh3wnx5PJKmjF3iQ6D5cO0he2GURiPeXlA4 uh87TJPlHJbOY1Cxd+UIG1JRigiDwOtr9V65868Y0DNeKVppAjk5h+iQJDxsth/Wi8x/ sDwyBMn/kf5sitLHaQYWL27RlXsxcI3Jg684lppLZe7n5h33Q6tJdjn/jMm4YuvxXwJl ZGCj3JuJtyrRjtn1QoYk1OnpoXviqRWtW7ZsZW8VGuKRnSawWY+55UJG/FihuANr6haY H7BeEXD2CpihDWfW2ZBErIlyS1JmYDqYIUIR9NI/NCzutsPrhhuEMsoFo/YTT5S2myM/ cFow== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id k4-v6si15687065pfc.328.2018.08.27.10.07.07; Mon, 27 Aug 2018 10:07:25 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727021AbeH0UxN (ORCPT + 99 others); Mon, 27 Aug 2018 16:53:13 -0400 Received: from mx3-rdu2.redhat.com ([66.187.233.73]:47354 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726809AbeH0UxN (ORCPT ); Mon, 27 Aug 2018 16:53:13 -0400 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 32ACB4023ECB; Mon, 27 Aug 2018 17:05:47 +0000 (UTC) Received: from krava (ovpn-204-43.brq.redhat.com [10.40.204.43]) by smtp.corp.redhat.com (Postfix) with SMTP id 6E7452166B41; Mon, 27 Aug 2018 17:05:45 +0000 (UTC) Date: Mon, 27 Aug 2018 19:05:43 +0200 From: Jiri Olsa To: Stephane Eranian Cc: LKML , Arnaldo Carvalho de Melo , Peter Zijlstra , mingo@elte.hu, Namhyung Kim Subject: [PATCHv3] perf tools: Add struct ordered_events_buffer layer Message-ID: <20180827170543.GA31347@krava> References: <1533767600-7794-1-git-send-email-eranian@google.com> <20180809080721.GB19243@krava> <20180810115431.GA4162@krava> <20180813130446.GA8685@krava> <20180815084825.GD3180@krava> <20180827092818.GA3725@krava> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.10.1 (2018-07-13) X-Scanned-By: MIMEDefang 2.78 on 10.11.54.6 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.7]); Mon, 27 Aug 2018 17:05:47 +0000 (UTC) X-Greylist: inspected by milter-greylist-4.5.16 (mx1.redhat.com [10.11.55.7]); Mon, 27 Aug 2018 17:05:47 +0000 (UTC) for IP:'10.11.54.6' DOMAIN:'int-mx06.intmail.prod.int.rdu2.redhat.com' HELO:'smtp.corp.redhat.com' FROM:'jolsa@redhat.com' RCPT:'' Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Aug 27, 2018 at 08:24:56AM -0700, Stephane Eranian wrote: SNIP > > - /* First entry is abused to maintain the to_free list. */ > > - oe->buffer_idx = 2; > > - new = oe->buffer + 1; > > + oe->buffer_idx = 1; > > + new = &oe->buffer->event[0]; > > } else { > > pr("allocation limit reached %" PRIu64 "B\n", oe->max_alloc_size); > > > I am wondering about the usefulness of returning a new_event with > new_event->event = NULL > in this case. Don't you need new_event->event? If so, then you need return NULL. yep, that's a bug.. with new being NULL in here, we'd get a crash anyway.. so 'return NULL;' it is SNIP > > + * yet, we need to free only allocated ones ... > > + */ > > + list_del(&oe->buffer->list); > > + ordered_events_buffer__free(oe->buffer, oe->buffer_idx, oe); > > + > > + /* ... and continue with the rest */ > > + list_for_each_entry_safe(buffer, tmp, &oe->to_free, list) { > > + list_del(&buffer->list); > > + ordered_events_buffer__free(buffer, MAX_SAMPLE_BUFFER, oe); > > > Here you are saying that if it is on the to_free list and not the > current buffer, then necessarily > all the entries have been used and it is safe to use > MAX_SAMPLE_BUFFER. Is that right? yes, at this point they either holds an event or NULL so it's free to call __free_dup_event on it thanks, v3 attached added also Namhyung's ack, as the 'return NULL' change wasn't related to the v2 changes jirka --- When ordering events, we use preallocated buffers to store separated events. Those buffers currently don't have their own struct, but since they are basically array of 'struct ordered_event' objects, we use the first event to hold buffers data - list head, that holds all buffers together: struct ordered_events { ... struct ordered_event *buffer; ... }; struct ordered_event { u64 timestamp; u64 file_offset; union perf_event *event; struct list_head list; }; This is quite convoluted and error prone as demonstrated by free-ing issue discovered and fixed by Stephane in here [1]. This patch adds the 'struct ordered_events_buffer' object, that holds the buffer data and frees it up properly. [1] - https://marc.info/?l=linux-kernel&m=153376761329335&w=2 Reported-by: Stephane Eranian Acked-by: Namhyung Kim Link: http://lkml.kernel.org/n/tip-qrkcqm5m1sugy4q83pfn5a1r@git.kernel.org Signed-off-by: Jiri Olsa --- tools/perf/util/ordered-events.c | 83 +++++++++++++++++++++++++++----- tools/perf/util/ordered-events.h | 37 ++++++++------ 2 files changed, 91 insertions(+), 29 deletions(-) diff --git a/tools/perf/util/ordered-events.c b/tools/perf/util/ordered-events.c index bad9e0296e9a..87171e8fd70d 100644 --- a/tools/perf/util/ordered-events.c +++ b/tools/perf/util/ordered-events.c @@ -80,14 +80,20 @@ static union perf_event *dup_event(struct ordered_events *oe, return oe->copy_on_queue ? __dup_event(oe, event) : event; } -static void free_dup_event(struct ordered_events *oe, union perf_event *event) +static void __free_dup_event(struct ordered_events *oe, union perf_event *event) { - if (event && oe->copy_on_queue) { + if (event) { oe->cur_alloc_size -= event->header.size; free(event); } } +static void free_dup_event(struct ordered_events *oe, union perf_event *event) +{ + if (oe->copy_on_queue) + __free_dup_event(oe, event); +} + #define MAX_SAMPLE_BUFFER (64 * 1024 / sizeof(struct ordered_event)) static struct ordered_event *alloc_event(struct ordered_events *oe, union perf_event *event) @@ -100,15 +106,43 @@ static struct ordered_event *alloc_event(struct ordered_events *oe, if (!new_event) return NULL; + /* + * We maintain following scheme of buffers for ordered + * event allocation: + * + * to_free list -> buffer1 (64K) + * buffer2 (64K) + * ... + * + * Each buffer keeps an array of ordered events objects: + * buffer -> event[0] + * event[1] + * ... + * + * Each allocated ordered event is linked to one of + * following lists: + * - time ordered list 'events' + * - list of currently removed events 'cache' + * + * Allocation of the ordered event uses following order + * to get the memory: + * - use recently removed object from 'cache' list + * - use available object in current allocation buffer + * - allocate new buffer if the current buffer is full + * + * Removal of ordered event object moves it from events to + * the cache list. + */ if (!list_empty(cache)) { new = list_entry(cache->next, struct ordered_event, list); list_del(&new->list); } else if (oe->buffer) { - new = oe->buffer + oe->buffer_idx; + new = &oe->buffer->event[oe->buffer_idx]; if (++oe->buffer_idx == MAX_SAMPLE_BUFFER) oe->buffer = NULL; } else if (oe->cur_alloc_size < oe->max_alloc_size) { - size_t size = MAX_SAMPLE_BUFFER * sizeof(*new); + size_t size = sizeof(*oe->buffer) + + MAX_SAMPLE_BUFFER * sizeof(*new); oe->buffer = malloc(size); if (!oe->buffer) { @@ -122,11 +156,11 @@ static struct ordered_event *alloc_event(struct ordered_events *oe, oe->cur_alloc_size += size; list_add(&oe->buffer->list, &oe->to_free); - /* First entry is abused to maintain the to_free list. */ - oe->buffer_idx = 2; - new = oe->buffer + 1; + oe->buffer_idx = 1; + new = &oe->buffer->event[0]; } else { pr("allocation limit reached %" PRIu64 "B\n", oe->max_alloc_size); + return NULL; } new->event = new_event; @@ -300,15 +334,38 @@ void ordered_events__init(struct ordered_events *oe, ordered_events__deliver_t d oe->deliver = deliver; } +static void +ordered_events_buffer__free(struct ordered_events_buffer *buffer, + unsigned int max, struct ordered_events *oe) +{ + if (oe->copy_on_queue) { + unsigned int i; + + for (i = 0; i < max; i++) + __free_dup_event(oe, buffer->event[i].event); + } + + free(buffer); +} + void ordered_events__free(struct ordered_events *oe) { - while (!list_empty(&oe->to_free)) { - struct ordered_event *event; + struct ordered_events_buffer *buffer, *tmp; - event = list_entry(oe->to_free.next, struct ordered_event, list); - list_del(&event->list); - free_dup_event(oe, event->event); - free(event); + if (list_empty(&oe->to_free)) + return; + + /* + * Current buffer might not have all the events allocated + * yet, we need to free only allocated ones ... + */ + list_del(&oe->buffer->list); + ordered_events_buffer__free(oe->buffer, oe->buffer_idx, oe); + + /* ... and continue with the rest */ + list_for_each_entry_safe(buffer, tmp, &oe->to_free, list) { + list_del(&buffer->list); + ordered_events_buffer__free(buffer, MAX_SAMPLE_BUFFER, oe); } } diff --git a/tools/perf/util/ordered-events.h b/tools/perf/util/ordered-events.h index 8c7a2948593e..1338d5c345dc 100644 --- a/tools/perf/util/ordered-events.h +++ b/tools/perf/util/ordered-events.h @@ -25,23 +25,28 @@ struct ordered_events; typedef int (*ordered_events__deliver_t)(struct ordered_events *oe, struct ordered_event *event); +struct ordered_events_buffer { + struct list_head list; + struct ordered_event event[0]; +}; + struct ordered_events { - u64 last_flush; - u64 next_flush; - u64 max_timestamp; - u64 max_alloc_size; - u64 cur_alloc_size; - struct list_head events; - struct list_head cache; - struct list_head to_free; - struct ordered_event *buffer; - struct ordered_event *last; - ordered_events__deliver_t deliver; - int buffer_idx; - unsigned int nr_events; - enum oe_flush last_flush_type; - u32 nr_unordered_events; - bool copy_on_queue; + u64 last_flush; + u64 next_flush; + u64 max_timestamp; + u64 max_alloc_size; + u64 cur_alloc_size; + struct list_head events; + struct list_head cache; + struct list_head to_free; + struct ordered_events_buffer *buffer; + struct ordered_event *last; + ordered_events__deliver_t deliver; + int buffer_idx; + unsigned int nr_events; + enum oe_flush last_flush_type; + u32 nr_unordered_events; + bool copy_on_queue; }; int ordered_events__queue(struct ordered_events *oe, union perf_event *event, -- 2.17.1