From patchwork Tue Dec 19 18:45:54 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Vincent Donnefort X-Patchwork-Id: 18322 Return-Path: Delivered-To: ouuuleilei@gmail.com Received: by 2002:a05:7300:24d3:b0:fb:cd0c:d3e with SMTP id r19csp2150183dyi; Tue, 19 Dec 2023 10:46:24 -0800 (PST) X-Google-Smtp-Source: AGHT+IE5A8oSjKGBegnLJAXJVPMoHLrfm7Pj0JrtyuUe2sCK+yZ9iNTKrAabaQEuJJyGOQT7Onxq X-Received: by 2002:a05:620a:248c:b0:780:e5d6:eb38 with SMTP id i12-20020a05620a248c00b00780e5d6eb38mr4986795qkn.86.1703011584385; Tue, 19 Dec 2023 10:46:24 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1703011584; cv=none; d=google.com; s=arc-20160816; b=hrgMuqV7cdoXKC6ABOg7kAoEkq56N/I3dViXTMHkirW2BV3BN9zJ8vBpxZWB5Sw5Bf YWCje3NNruboBKGqbxg1DW8lo5qcSzDL/3mf7/V9v0F4UTjwL3QqChkNBevN8RsQ6g9N IkEETzaySN4yq/f42SOdXa8cBzn8j+Hp+aHGpplSRGjE5tIyZO+YzxUDdNZiOlgRnYL9 xuECHwKRM3KLUfleQZK+TymKHyaIActareY4I8Yn0ENf/vaT2tyBU9vea+i/VoXzkoNZ 9j2aBpcxibqR5qI9KVzp8gmKP7jTiqtvqdo3mGPQGip/BeHYkEs5tffDTexQIjNZDDsu cnDw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=cc:to:from:subject:message-id:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:dkim-signature; bh=KLOEeGfb/XH+b9fpUeX+22oz7bMV4sUzfwYQZztEShI=; fh=BzJUDniL4L4xFu6YvrY/01a7DDZvz+Mdyin6gi665+Y=; b=C7OpPF0hFwtj+b82hgF5A++GYdnUrQ5JWRQ8haBMbeMbEKkN0X2tGuPvSgxtTsCV6F T5EvF38RkUaGnelsKOfn7Bt8pYX9dxfcDBfEO2GgIjnEgqOBaE3Ms2HZEpJ80WHMxs9k CaVeP4nmlGZTcltVWXqKbnyIw2w8/NmTICVWMtA1NgKM48EBlalXpT+1XtycHuiWuhLy RdV+ZfL4wG6+yKhvfEiCG15G95xFYkApB6qs379EZ53KFOF/hzIGiGexP2QCSrlDAnkz zQTpDT4HY+ozunZHUhrtIZ6ME9haVtkC7F463HM0nrZWmGJcOyt+yIMRzF5wZT6Nyoh3 X+Gg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=0FNL6BXa; spf=pass (google.com: domain of linux-kernel+bounces-5824-ouuuleilei=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-5824-ouuuleilei=gmail.com@vger.kernel.org"; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [147.75.199.223]) by mx.google.com with ESMTPS id s9-20020a05620a030900b007810bc85f47si621635qkm.599.2023.12.19.10.46.24 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 19 Dec 2023 10:46:24 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-5824-ouuuleilei=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) client-ip=147.75.199.223; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=0FNL6BXa; spf=pass (google.com: domain of linux-kernel+bounces-5824-ouuuleilei=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-5824-ouuuleilei=gmail.com@vger.kernel.org"; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 1FCC01C21879 for ; Tue, 19 Dec 2023 18:46:24 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 1B82838FA8; Tue, 19 Dec 2023 18:46:07 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=google.com header.i=@google.com header.b="0FNL6BXa" X-Original-To: linux-kernel@vger.kernel.org Received: from mail-wm1-f74.google.com (mail-wm1-f74.google.com [209.85.128.74]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 9989A38DEE for ; Tue, 19 Dec 2023 18:46:02 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=reject dis=none) header.from=google.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=flex--vdonnefort.bounces.google.com Received: by mail-wm1-f74.google.com with SMTP id 5b1f17b1804b1-40c27a351e2so32971785e9.2 for ; Tue, 19 Dec 2023 10:46:02 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1703011561; x=1703616361; darn=vger.kernel.org; h=cc:to:from:subject:message-id:mime-version:date:from:to:cc:subject :date:message-id:reply-to; bh=KLOEeGfb/XH+b9fpUeX+22oz7bMV4sUzfwYQZztEShI=; b=0FNL6BXah+geCnCMQnGBDzzBAvKKeXPcx2ingsuIHoI5Rp5Sic6JF6hmPL21sw9lAj HLTI0+H32svrzovLfsWbhD0sMUZ4+ps0Kysq76n5Vp3X9kqJf0Fo9eO+MOBKiYhqXOwJ FCovW2/cGUXFffdL8H4uYuoMJIYceBYfBf8v87ayQ2lxR9jZ2f8MVYa0K34+BLFJlweQ SQHsaUMCHd7TRxt5AwV+jiXVzzm+bBN49UZ1dge6RiECoEdEUmiLPUniRe4vcPO1EJEr uPd7aTcTYEQ26iGP8dZ6c4sgw4ABnZ2s4tM3TLQ7a2Zza0o564HSI3p0Ilqn7sm5El8y FKwg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1703011561; x=1703616361; h=cc:to:from:subject:message-id:mime-version:date:x-gm-message-state :from:to:cc:subject:date:message-id:reply-to; bh=KLOEeGfb/XH+b9fpUeX+22oz7bMV4sUzfwYQZztEShI=; b=jBCtDob+AFcmvx9fa8IAr5VKnQ8GOHgyG+D0nkkl2odF/p+QAfWBp/qGxoDHN35TQH vv1fZQMAQmZF69D5ggzS8sR2b2nS2d3BuDLnSA8E5AulQWaYEYoalILheQ8clygk2911 2coocnGem7BfG06+QtLxE1FbfmzyHjbLsSV/fE2H+PVuHzNovNtx3AOd6etBZRAOh4Ta wd72kX8C4oNZNjSEr2NE6vre/ktRG1fvUDiltwPQJr2Z0/3lx62xd53uhbjlaIyKG7uL hB8E/gd8QkChWkJgdExNDF0UB63rL8gryLdS9o/I3LSp0hdbxZey97dXP/+GAh92MuDM EtWw== X-Gm-Message-State: AOJu0Yz8Qi+bnOskMMh3QRQ9mG0VYBFjTc8zaoggBPxoqrRWPakYShAk HIfMRiICBzWSNwIlSjVxUtHg5S45lgZBVvTo X-Received: from vdonnefort.c.googlers.com ([fda3:e722:ac3:cc00:28:9cb1:c0a8:2eea]) (user=vdonnefort job=sendgmr) by 2002:a05:600c:3b82:b0:40d:2021:9c27 with SMTP id n2-20020a05600c3b8200b0040d20219c27mr14246wms.6.1703011560651; Tue, 19 Dec 2023 10:46:00 -0800 (PST) Date: Tue, 19 Dec 2023 18:45:54 +0000 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: Mime-Version: 1.0 X-Mailer: git-send-email 2.43.0.472.g3155946c3a-goog Message-ID: <20231219184556.1552951-1-vdonnefort@google.com> Subject: [PATCH v8 0/2] Introducing trace buffer mapping by user-space From: Vincent Donnefort To: rostedt@goodmis.org, mhiramat@kernel.org, linux-kernel@vger.kernel.org, linux-trace-kernel@vger.kernel.org Cc: kernel-team@android.com, Vincent Donnefort X-getmail-retrieved-from-mailbox: INBOX X-GMAIL-THRID: 1785737075154846682 X-GMAIL-MSGID: 1785737075154846682 The tracing ring-buffers can be stored on disk or sent to network without any copy via splice. However the later doesn't allow real time processing of the traces. A solution is to give userspace direct access to the ring-buffer pages via a mapping. An application can now become a consumer of the ring-buffer, in a similar fashion to what trace_pipe offers. Attached to this cover letter an example of consuming read for a ring-buffer, using libtracefs. Vincent v7 -> v8: * Drop the subbufs renaming into bpages * Use subbuf as a name when relevant v6 -> v7: * Rebase onto lore.kernel.org/lkml/20231215175502.106587604@goodmis.org/ * Support for subbufs * Rename subbufs into bpages v5 -> v6: * Rebase on next-20230802. * (unsigned long) -> (void *) cast for virt_to_page(). * Add a wait for the GET_READER_PAGE ioctl. * Move writer fields update (overrun/pages_lost/entries/pages_touched) in the irq_work. * Rearrange id in struct buffer_page. * Rearrange the meta-page. * ring_buffer_meta_page -> trace_buffer_meta_page. * Add meta_struct_len into the meta-page. v4 -> v5: * Trivial rebase onto 6.5-rc3 (previously 6.4-rc3) v3 -> v4: * Add to the meta-page: - pages_lost / pages_read (allow to compute how full is the ring-buffer) - read (allow to compute how many entries can be read) - A reader_page struct. * Rename ring_buffer_meta_header -> ring_buffer_meta * Rename ring_buffer_get_reader_page -> ring_buffer_map_get_reader_page * Properly consume events on ring_buffer_map_get_reader_page() with rb_advance_reader(). v2 -> v3: * Remove data page list (for non-consuming read) ** Implies removing order > 0 meta-page * Add a new meta page field ->read * Rename ring_buffer_meta_page_header into ring_buffer_meta_header v1 -> v2: * Hide data_pages from the userspace struct * Fix META_PAGE_MAX_PAGES * Support for order > 0 meta-page * Add missing page->mapping. --- /* Need to access private struct to save counters */ struct kbuffer { unsigned long long timestamp; long long lost_events; unsigned long flags; void *subbuffer; void *data; unsigned int index; unsigned int curr; unsigned int next; unsigned int size; unsigned int start; unsigned int first; unsigned int (*read_4)(void *ptr); unsigned long long (*read_8)(void *ptr); unsigned long long (*read_long)(struct kbuffer *kbuf, void *ptr); int (*next_event)(struct kbuffer *kbuf); }; struct trace_buffer_meta { unsigned long entries; unsigned long overrun; unsigned long read; unsigned long subbufs_touched; unsigned long subbufs_lost; unsigned long subbufs_read; struct { unsigned long lost_events; /* Events lost at the time of the reader swap */ __u32 id; /* Reader subbuf ID from 0 to nr_subbufs - 1 */ __u32 read; /* Number of bytes read on the reader subbuf */ } reader; __u32 subbuf_size; __u32 nr_subbufs; /* Number of subbufs in the ring-buffer */ __u32 meta_page_size; /* Size of the meta-page */ __u32 meta_struct_len; /* Len of this struct */ }; static char *argv0; static bool exit_requested; static char *get_this_name(void) { static char *this_name; char *arg; char *p; if (this_name) return this_name; arg = argv0; p = arg+strlen(arg); while (p >= arg && *p != '/') p--; p++; this_name = p; return p; } static void __vdie(const char *fmt, va_list ap, int err) { int ret = errno; char *p = get_this_name(); if (err && errno) perror(p); else ret = -1; fprintf(stderr, " "); vfprintf(stderr, fmt, ap); fprintf(stderr, "\n"); exit(ret); } void pdie(const char *fmt, ...) { va_list ap; va_start(ap, fmt); __vdie(fmt, ap, 1); va_end(ap); } static void read_subbuf(struct tep_handle *tep, struct kbuffer *kbuf) { static struct trace_seq seq; struct tep_record record; if (seq.buffer) trace_seq_reset(&seq); else trace_seq_init(&seq); while ((record.data = kbuffer_read_event(kbuf, &record.ts))) { record.size = kbuffer_event_size(kbuf); kbuffer_next_event(kbuf, NULL); tep_print_event(tep, &seq, &record, "%s-%d %9d\t%s: %s\n", TEP_PRINT_COMM, TEP_PRINT_PID, TEP_PRINT_TIME, TEP_PRINT_NAME, TEP_PRINT_INFO); trace_seq_do_printf(&seq); trace_seq_reset(&seq); } } static int next_reader_subbuf(int fd, struct trace_buffer_meta *meta, unsigned long *read) { __u32 prev_read, prev_reader, new_reader; prev_read = READ_ONCE(meta->reader.read); prev_reader = READ_ONCE(meta->reader.id); if (ioctl(fd, TRACE_MMAP_IOCTL_GET_READER) < 0) pdie("ioctl"); new_reader = READ_ONCE(meta->reader.id); if (prev_reader != new_reader) *read = 0; else *read = prev_read; return new_reader; } static void signal_handler(int unused) { printf("Exit!\n"); exit_requested = true; } int main(int argc, char **argv) { int page_size, meta_len, data_len, subbuf, fd; struct trace_buffer_meta *map; struct tep_handle *tep; struct kbuffer *kbuf; unsigned long read; void *meta, *data; char path[32]; int cpu; if (argc != 2) return -EINVAL; argv0 = argv[0]; cpu = atoi(argv[1]); snprintf(path, 32, "per_cpu/cpu%d/trace_pipe_raw", cpu); tep = tracefs_local_events(NULL); kbuf = tep_kbuffer(tep); page_size = getpagesize(); fd = tracefs_instance_file_open(NULL, path, O_RDONLY); if (fd < 0) pdie("raw"); meta = mmap(NULL, page_size, PROT_READ, MAP_SHARED, fd, 0); if (meta == MAP_FAILED) pdie("mmap"); map = (struct trace_buffer_meta *)meta; meta_len = map->meta_page_size; printf("entries: %lu\n", map->entries); printf("overrun: %lu\n", map->overrun); printf("read: %lu\n", map->read); printf("subbufs_touched:%lu\n", map->subbufs_touched); printf("subbufs_lost: %lu\n", map->subbufs_lost); printf("subbufs_read: %lu\n", map->subbufs_read); printf("nr_subbufs: %u\n", map->nr_subbufs); data_len = map->subbuf_size * map->nr_subbufs; data = mmap(NULL, data_len, PROT_READ, MAP_SHARED, fd, meta_len); if (data == MAP_FAILED) pdie("mmap data"); signal(SIGINT, signal_handler); while (!exit_requested) { subbuf = next_reader_subbuf(fd, map, &read); kbuffer_load_subbuffer(kbuf, data + map->subbuf_size * subbuf); while (kbuf->curr < read) kbuffer_next_event(kbuf, NULL); read_subbuf(tep, kbuf); } munmap(data, data_len); munmap(meta, page_size); close(fd); return 0; } Vincent Donnefort (2): ring-buffer: Introducing ring-buffer mapping functions tracing: Allow user-space mapping of the ring-buffer include/linux/ring_buffer.h | 7 + include/uapi/linux/trace_mmap.h | 31 +++ kernel/trace/ring_buffer.c | 371 +++++++++++++++++++++++++++++++- kernel/trace/trace.c | 79 ++++++- 4 files changed, 484 insertions(+), 4 deletions(-) create mode 100644 include/uapi/linux/trace_mmap.h