X-Git-Url: http://git.lttng.org./?a=blobdiff_plain;f=wrapper%2Fvmalloc.h;h=75c7e2c66519611a1ecfd027f44ef992048d971e;hb=4ad901836eb32be6119bff1313a23cfb602218b1;hp=b4d795404a2dfa454f9b1dae11a0f0d8963db6c9;hpb=20eb87c94b862d82c5c5eb279a2dc085045059b5;p=lttng-modules.git diff --git a/wrapper/vmalloc.h b/wrapper/vmalloc.h index b4d79540..75c7e2c6 100644 --- a/wrapper/vmalloc.h +++ b/wrapper/vmalloc.h @@ -1,7 +1,5 @@ -#ifndef _LTTNG_WRAPPER_VMALLOC_H -#define _LTTNG_WRAPPER_VMALLOC_H - -/* +/* SPDX-License-Identifier: (GPL-2.0 or LGPL-2.1) + * * wrapper/vmalloc.h * * wrapper around vmalloc_sync_all. Using KALLSYMS to get its address when @@ -9,22 +7,11 @@ * modules. * * Copyright (C) 2011-2012 Mathieu Desnoyers - * - * This library is free software; you can redistribute it and/or - * modify it under the terms of the GNU Lesser General Public - * License as published by the Free Software Foundation; only - * version 2.1 of the License. - * - * This library is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - * Lesser General Public License for more details. - * - * You should have received a copy of the GNU Lesser General Public - * License along with this library; if not, write to the Free Software - * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ +#ifndef _LTTNG_WRAPPER_VMALLOC_H +#define _LTTNG_WRAPPER_VMALLOC_H + #include #include #include @@ -33,9 +20,46 @@ #include #include +#include + +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ + || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ + || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ + || LTTNG_KERNEL_RANGE(4,19,113, 4,20,0) \ + || LTTNG_KERNEL_RANGE(4,14,175, 4,15,0) \ + || LTTNG_KERNEL_RANGE(4,9,218, 4,10,0) \ + || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0)) \ + || LTTNG_UBUNTU_KERNEL_RANGE(4,15,18,97, 4,16,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,0,21,48, 5,1,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) + +static inline +void wrapper_vmalloc_sync_mappings(void) +{ + void (*vmalloc_sync_mappings_sym)(void); + + vmalloc_sync_mappings_sym = (void *) kallsyms_lookup_funcptr("vmalloc_sync_mappings"); + if (vmalloc_sync_mappings_sym) { + vmalloc_sync_mappings_sym(); + } else { +#ifdef CONFIG_X86 + /* + * Only x86 needs vmalloc_sync_mappings to make sure LTTng does not + * trigger recursive page faults. + */ + printk_once(KERN_WARNING "LTTng: vmalloc_sync_mappings symbol lookup failed.\n"); + printk_once(KERN_WARNING "Page fault handler and NMI tracing might trigger faults.\n"); +#endif + } +} + +#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ +/* + * Map vmalloc_sync_mappings to vmalloc_sync_all() on kernels before 5.7. + */ static inline -void wrapper_vmalloc_sync_all(void) +void wrapper_vmalloc_sync_mappings(void) { void (*vmalloc_sync_all_sym)(void); @@ -53,13 +77,38 @@ void wrapper_vmalloc_sync_all(void) #endif } } + +#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ + #else +#if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0) \ + || LTTNG_KERNEL_RANGE(5,4,28, 5,5,0) \ + || LTTNG_KERNEL_RANGE(5,2,37, 5,3,0) \ + || LTTNG_KERNEL_RANGE(4,19,113, 4,20,0) \ + || LTTNG_KERNEL_RANGE(4,14,175, 4,15,0) \ + || LTTNG_KERNEL_RANGE(4,9,218, 4,10,0) \ + || LTTNG_KERNEL_RANGE(4,4,218, 4,5,0)) \ + || LTTNG_UBUNTU_KERNEL_RANGE(4,15,18,97, 4,18,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,0,21,48, 5,1,0,0) \ + || LTTNG_UBUNTU_KERNEL_RANGE(5,3,18,52, 5,4,0,0) + static inline -void wrapper_vmalloc_sync_all(void) +void wrapper_vmalloc_sync_mappings(void) +{ + return vmalloc_sync_mappings(); +} + +#else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ + +static inline +void wrapper_vmalloc_sync_mappings(void) { return vmalloc_sync_all(); } + +#endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(5,6,0)) */ + #endif #if (LINUX_VERSION_CODE >= KERNEL_VERSION(4,12,0)) @@ -74,7 +123,7 @@ void *lttng_kvmalloc_node(unsigned long size, gfp_t flags, int node) * Make sure we don't trigger recursive page faults in the * tracing fast path. */ - wrapper_vmalloc_sync_all(); + wrapper_vmalloc_sync_mappings(); } return ret; } @@ -186,7 +235,7 @@ void *lttng_kvmalloc_node(unsigned long size, gfp_t flags, int node) * Make sure we don't trigger recursive page faults in the * tracing fast path. */ - wrapper_vmalloc_sync_all(); + wrapper_vmalloc_sync_mappings(); } return ret; }