Skip to content

Commit 2dba5eb

Browse files
tehcastertorvalds
authored andcommitted
lib/stackdepot: allow optional init and stack_table allocation by kvmalloc()
Currently, enabling CONFIG_STACKDEPOT means its stack_table will be allocated from memblock, even if stack depot ends up not actually used. The default size of stack_table is 4MB on 32-bit, 8MB on 64-bit. This is fine for use-cases such as KASAN which is also a config option and has overhead on its own. But it's an issue for functionality that has to be actually enabled on boot (page_owner) or depends on hardware (GPU drivers) and thus the memory might be wasted. This was raised as an issue [1] when attempting to add stackdepot support for SLUB's debug object tracking functionality. It's common to build kernels with CONFIG_SLUB_DEBUG and enable slub_debug on boot only when needed, or create only specific kmem caches with debugging for testing purposes. It would thus be more efficient if stackdepot's table was allocated only when actually going to be used. This patch thus makes the allocation (and whole stack_depot_init() call) optional: - Add a CONFIG_STACKDEPOT_ALWAYS_INIT flag to keep using the current well-defined point of allocation as part of mem_init(). Make CONFIG_KASAN select this flag. - Other users have to call stack_depot_init() as part of their own init when it's determined that stack depot will actually be used. This may depend on both config and runtime conditions. Convert current users which are page_owner and several in the DRM subsystem. Same will be done for SLUB later. - Because the init might now be called after the boot-time memblock allocation has given all memory to the buddy allocator, change stack_depot_init() to allocate stack_table with kvmalloc() when memblock is no longer available. Also handle allocation failure by disabling stackdepot (could have theoretically happened even with memblock allocation previously), and don't unnecessarily align the memblock allocation to its own size anymore. [1] https://lore.kernel.org/all/CAMuHMdW=eoVzM1Re5FVoEN87nKfiLmM2+Ah7eNu2KXEhCvbZyA@mail.gmail.com/ Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Vlastimil Babka <[email protected]> Acked-by: Dmitry Vyukov <[email protected]> Reviewed-by: Marco Elver <[email protected]> # stackdepot Cc: Marco Elver <[email protected]> Cc: Vijayanand Jitta <[email protected]> Cc: Maarten Lankhorst <[email protected]> Cc: Maxime Ripard <[email protected]> Cc: Thomas Zimmermann <[email protected]> Cc: David Airlie <[email protected]> Cc: Daniel Vetter <[email protected]> Cc: Andrey Ryabinin <[email protected]> Cc: Alexander Potapenko <[email protected]> Cc: Andrey Konovalov <[email protected]> Cc: Dmitry Vyukov <[email protected]> Cc: Geert Uytterhoeven <[email protected]> Cc: Oliver Glitta <[email protected]> Cc: Imran Khan <[email protected]> From: Colin Ian King <[email protected]> Subject: lib/stackdepot: fix spelling mistake and grammar in pr_err message There is a spelling mistake of the work allocation so fix this and re-phrase the message to make it easier to read. Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Colin Ian King <[email protected]> Cc: Vlastimil Babka <[email protected]> From: Vlastimil Babka <[email protected]> Subject: lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup On FLATMEM, we call page_ext_init_flatmem_late() just before kmem_cache_init() which means stack_depot_init() (called by page owner init) will not recognize properly it should use kvmalloc() and not memblock_alloc(). memblock_alloc() will also not issue a warning and return a block memory that can be invalid and cause kernel page fault when saving stacks, as reported by the kernel test robot [1]. Fix this by moving page_ext_init_flatmem_late() below kmem_cache_init() so that slab_is_available() is true during stack_depot_init(). SPARSEMEM doesn't have this issue, as it doesn't do page_ext_init_flatmem_late(), but a different page_ext_init() even later in the boot process. Thanks to Mike Rapoport for pointing out the FLATMEM init ordering issue. While at it, also actually resolve a checkpatch warning in stack_depot_init() from DRM CI, which was supposed to be in the original patch already. [1] https://lore.kernel.org/all/20211014085450.GC18719@xsang-OptiPlex-9020/ Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Vlastimil Babka <[email protected]> Reported-by: kernel test robot <[email protected]> Cc: Mike Rapoport <[email protected]> Cc: Stephen Rothwell <[email protected]> From: Vlastimil Babka <[email protected]> Subject: lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup3 Due to cd06ab2 ("drm/locking: add backtrace for locking contended locks without backoff") landing recently to -next adding a new stack depot user in drivers/gpu/drm/drm_modeset_lock.c we need to add an appropriate call to stack_depot_init() there as well. Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Vlastimil Babka <[email protected]> Cc: Jani Nikula <[email protected]> Cc: Naresh Kamboju <[email protected]> Cc: Marco Elver <[email protected]> Cc: Vijayanand Jitta <[email protected]> Cc: Maarten Lankhorst <[email protected]> Cc: Maxime Ripard <[email protected]> Cc: Thomas Zimmermann <[email protected]> Cc: David Airlie <[email protected]> Cc: Daniel Vetter <[email protected]> Cc: Andrey Ryabinin <[email protected]> Cc: Alexander Potapenko <[email protected]> Cc: Andrey Konovalov <[email protected]> Cc: Dmitry Vyukov <[email protected]> Cc: Geert Uytterhoeven <[email protected]> Cc: Oliver Glitta <[email protected]> Cc: Imran Khan <[email protected]> Cc: Stephen Rothwell <[email protected]> From: Vlastimil Babka <[email protected]> Subject: lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup4 Due to 4e66934 ("lib: add reference counting tracking infrastructure") landing recently to net-next adding a new stack depot user in lib/ref_tracker.c we need to add an appropriate call to stack_depot_init() there as well. Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Vlastimil Babka <[email protected]> Reviewed-by: Eric Dumazet <[email protected]> Cc: Jiri Slab <[email protected]> Signed-off-by: Andrew Morton <[email protected]> Signed-off-by: Linus Torvalds <[email protected]>
1 parent 359745d commit 2dba5eb

File tree

11 files changed

+76
-18
lines changed

11 files changed

+76
-18
lines changed

drivers/gpu/drm/drm_dp_mst_topology.c

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5511,6 +5511,7 @@ int drm_dp_mst_topology_mgr_init(struct drm_dp_mst_topology_mgr *mgr,
55115511
mutex_init(&mgr->probe_lock);
55125512
#if IS_ENABLED(CONFIG_DRM_DEBUG_DP_MST_TOPOLOGY_REFS)
55135513
mutex_init(&mgr->topology_ref_history_lock);
5514+
stack_depot_init();
55145515
#endif
55155516
INIT_LIST_HEAD(&mgr->tx_msg_downq);
55165517
INIT_LIST_HEAD(&mgr->destroy_port_list);

drivers/gpu/drm/drm_mm.c

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -980,6 +980,10 @@ void drm_mm_init(struct drm_mm *mm, u64 start, u64 size)
980980
add_hole(&mm->head_node);
981981

982982
mm->scan_active = 0;
983+
984+
#ifdef CONFIG_DRM_DEBUG_MM
985+
stack_depot_init();
986+
#endif
983987
}
984988
EXPORT_SYMBOL(drm_mm_init);
985989

drivers/gpu/drm/drm_modeset_lock.c

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -107,6 +107,11 @@ static void __drm_stack_depot_print(depot_stack_handle_t stack_depot)
107107

108108
kfree(buf);
109109
}
110+
111+
static void __drm_stack_depot_init(void)
112+
{
113+
stack_depot_init();
114+
}
110115
#else /* CONFIG_DRM_DEBUG_MODESET_LOCK */
111116
static depot_stack_handle_t __drm_stack_depot_save(void)
112117
{
@@ -115,6 +120,9 @@ static depot_stack_handle_t __drm_stack_depot_save(void)
115120
static void __drm_stack_depot_print(depot_stack_handle_t stack_depot)
116121
{
117122
}
123+
static void __drm_stack_depot_init(void)
124+
{
125+
}
118126
#endif /* CONFIG_DRM_DEBUG_MODESET_LOCK */
119127

120128
/**
@@ -359,6 +367,7 @@ void drm_modeset_lock_init(struct drm_modeset_lock *lock)
359367
{
360368
ww_mutex_init(&lock->mutex, &crtc_ww_class);
361369
INIT_LIST_HEAD(&lock->head);
370+
__drm_stack_depot_init();
362371
}
363372
EXPORT_SYMBOL(drm_modeset_lock_init);
364373

drivers/gpu/drm/i915/intel_runtime_pm.c

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -68,6 +68,9 @@ static noinline depot_stack_handle_t __save_depot_stack(void)
6868
static void init_intel_runtime_pm_wakeref(struct intel_runtime_pm *rpm)
6969
{
7070
spin_lock_init(&rpm->debug.lock);
71+
72+
if (rpm->available)
73+
stack_depot_init();
7174
}
7275

7376
static noinline depot_stack_handle_t

include/linux/ref_tracker.h

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@
44
#include <linux/refcount.h>
55
#include <linux/types.h>
66
#include <linux/spinlock.h>
7+
#include <linux/stackdepot.h>
78

89
struct ref_tracker;
910

@@ -26,6 +27,7 @@ static inline void ref_tracker_dir_init(struct ref_tracker_dir *dir,
2627
spin_lock_init(&dir->lock);
2728
dir->quarantine_avail = quarantine_count;
2829
refcount_set(&dir->untracked, 1);
30+
stack_depot_init();
2931
}
3032

3133
void ref_tracker_dir_exit(struct ref_tracker_dir *dir);

include/linux/stackdepot.h

Lines changed: 16 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,22 @@ depot_stack_handle_t __stack_depot_save(unsigned long *entries,
1919
unsigned int nr_entries,
2020
gfp_t gfp_flags, bool can_alloc);
2121

22+
/*
23+
* Every user of stack depot has to call this during its own init when it's
24+
* decided that it will be calling stack_depot_save() later.
25+
*
26+
* The alternative is to select STACKDEPOT_ALWAYS_INIT to have stack depot
27+
* enabled as part of mm_init(), for subsystems where it's known at compile time
28+
* that stack depot will be used.
29+
*/
30+
int stack_depot_init(void);
31+
32+
#ifdef CONFIG_STACKDEPOT_ALWAYS_INIT
33+
static inline int stack_depot_early_init(void) { return stack_depot_init(); }
34+
#else
35+
static inline int stack_depot_early_init(void) { return 0; }
36+
#endif
37+
2238
depot_stack_handle_t stack_depot_save(unsigned long *entries,
2339
unsigned int nr_entries, gfp_t gfp_flags);
2440

@@ -30,13 +46,4 @@ int stack_depot_snprint(depot_stack_handle_t handle, char *buf, size_t size,
3046

3147
void stack_depot_print(depot_stack_handle_t stack);
3248

33-
#ifdef CONFIG_STACKDEPOT
34-
int stack_depot_init(void);
35-
#else
36-
static inline int stack_depot_init(void)
37-
{
38-
return 0;
39-
}
40-
#endif /* CONFIG_STACKDEPOT */
41-
4249
#endif

init/main.c

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -834,12 +834,15 @@ static void __init mm_init(void)
834834
init_mem_debugging_and_hardening();
835835
kfence_alloc_pool();
836836
report_meminit();
837-
stack_depot_init();
837+
stack_depot_early_init();
838838
mem_init();
839839
mem_init_print_info();
840-
/* page_owner must be initialized after buddy is ready */
841-
page_ext_init_flatmem_late();
842840
kmem_cache_init();
841+
/*
842+
* page_owner must be initialized after buddy is ready, and also after
843+
* slab is ready so that stack_depot_init() works properly
844+
*/
845+
page_ext_init_flatmem_late();
843846
kmemleak_init();
844847
pgtable_init();
845848
debug_objects_mem_init();

lib/Kconfig

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -673,6 +673,10 @@ config STACKDEPOT
673673
bool
674674
select STACKTRACE
675675

676+
config STACKDEPOT_ALWAYS_INIT
677+
bool
678+
select STACKDEPOT
679+
676680
config STACK_HASH_ORDER
677681
int "stack depot hash size (12 => 4KB, 20 => 1024KB)"
678682
range 12 20

lib/Kconfig.kasan

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ menuconfig KASAN
3838
CC_HAS_WORKING_NOSANITIZE_ADDRESS) || \
3939
HAVE_ARCH_KASAN_HW_TAGS
4040
depends on (SLUB && SYSFS) || (SLAB && !DEBUG_SLAB)
41-
select STACKDEPOT
41+
select STACKDEPOT_ALWAYS_INIT
4242
help
4343
Enables KASAN (KernelAddressSANitizer) - runtime memory debugger,
4444
designed to find out-of-bounds accesses and use-after-free bugs.

lib/stackdepot.c

Lines changed: 28 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@
2323
#include <linux/jhash.h>
2424
#include <linux/kernel.h>
2525
#include <linux/mm.h>
26+
#include <linux/mutex.h>
2627
#include <linux/percpu.h>
2728
#include <linux/printk.h>
2829
#include <linux/slab.h>
@@ -161,18 +162,40 @@ static int __init is_stack_depot_disabled(char *str)
161162
}
162163
early_param("stack_depot_disable", is_stack_depot_disabled);
163164

164-
int __init stack_depot_init(void)
165+
/*
166+
* __ref because of memblock_alloc(), which will not be actually called after
167+
* the __init code is gone, because at that point slab_is_available() is true
168+
*/
169+
__ref int stack_depot_init(void)
165170
{
166-
if (!stack_depot_disable) {
171+
static DEFINE_MUTEX(stack_depot_init_mutex);
172+
173+
mutex_lock(&stack_depot_init_mutex);
174+
if (!stack_depot_disable && !stack_table) {
167175
size_t size = (STACK_HASH_SIZE * sizeof(struct stack_record *));
168176
int i;
169177

170-
stack_table = memblock_alloc(size, size);
171-
for (i = 0; i < STACK_HASH_SIZE; i++)
172-
stack_table[i] = NULL;
178+
if (slab_is_available()) {
179+
pr_info("Stack Depot allocating hash table with kvmalloc\n");
180+
stack_table = kvmalloc(size, GFP_KERNEL);
181+
} else {
182+
pr_info("Stack Depot allocating hash table with memblock_alloc\n");
183+
stack_table = memblock_alloc(size, SMP_CACHE_BYTES);
184+
}
185+
if (stack_table) {
186+
for (i = 0; i < STACK_HASH_SIZE; i++)
187+
stack_table[i] = NULL;
188+
} else {
189+
pr_err("Stack Depot hash table allocation failed, disabling\n");
190+
stack_depot_disable = true;
191+
mutex_unlock(&stack_depot_init_mutex);
192+
return -ENOMEM;
193+
}
173194
}
195+
mutex_unlock(&stack_depot_init_mutex);
174196
return 0;
175197
}
198+
EXPORT_SYMBOL_GPL(stack_depot_init);
176199

177200
/* Calculate hash for a stack */
178201
static inline u32 hash_stack(unsigned long *entries, unsigned int size)

0 commit comments

Comments
 (0)