kfence: fix stack trace pruning
[linux-2.6-microblaze.git] / mm / kfence / report.c
index 7e49685..46ecea1 100644 (file)
@@ -75,18 +75,23 @@ static int get_stack_skipnr(const unsigned long stack_entries[], int num_entries
 
                if (str_has_prefix(buf, ARCH_FUNC_PREFIX "kfence_") ||
                    str_has_prefix(buf, ARCH_FUNC_PREFIX "__kfence_") ||
+                   str_has_prefix(buf, ARCH_FUNC_PREFIX "__kmem_cache_free") ||
                    !strncmp(buf, ARCH_FUNC_PREFIX "__slab_free", len)) {
                        /*
-                        * In case of tail calls from any of the below
-                        * to any of the above.
+                        * In case of tail calls from any of the below to any of
+                        * the above, optimized by the compiler such that the
+                        * stack trace would omit the initial entry point below.
                         */
                        fallback = skipnr + 1;
                }
 
-               /* Also the *_bulk() variants by only checking prefixes. */
+               /*
+                * The below list should only include the initial entry points
+                * into the slab allocators. Includes the *_bulk() variants by
+                * checking prefixes.
+                */
                if (str_has_prefix(buf, ARCH_FUNC_PREFIX "kfree") ||
                    str_has_prefix(buf, ARCH_FUNC_PREFIX "kmem_cache_free") ||
-                   str_has_prefix(buf, ARCH_FUNC_PREFIX "__kmem_cache_free") ||
                    str_has_prefix(buf, ARCH_FUNC_PREFIX "__kmalloc") ||
                    str_has_prefix(buf, ARCH_FUNC_PREFIX "kmem_cache_alloc"))
                        goto found;