Commit d224eb02 authored by Suren Baghdasaryan's avatar Suren Baghdasaryan Committed by Andrew Morton

codetag: debug: mark codetags for reserved pages as empty

To avoid debug warnings while freeing reserved pages which were not
allocated with usual allocators, mark their codetags as empty before
freeing.

Link: https://lkml.kernel.org/r/20240321163705.3067592-35-surenb@google.comSigned-off-by: default avatarSuren Baghdasaryan <surenb@google.com>
Reviewed-by: default avatarKees Cook <keescook@chromium.org>
Tested-by: default avatarKees Cook <keescook@chromium.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Alex Gaynor <alex.gaynor@gmail.com>
Cc: Alice Ryhl <aliceryhl@google.com>
Cc: Andreas Hindborg <a.hindborg@samsung.com>
Cc: Benno Lossin <benno.lossin@proton.me>
Cc: "Björn Roy Baron" <bjorn3_gh@protonmail.com>
Cc: Boqun Feng <boqun.feng@gmail.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Gary Guo <gary@garyguo.net>
Cc: Kent Overstreet <kent.overstreet@linux.dev>
Cc: Miguel Ojeda <ojeda@kernel.org>
Cc: Pasha Tatashin <pasha.tatashin@soleen.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Wedson Almeida Filho <wedsonaf@gmail.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 239d6c96
...@@ -47,6 +47,7 @@ static inline void set_codetag_empty(union codetag_ref *ref) ...@@ -47,6 +47,7 @@ static inline void set_codetag_empty(union codetag_ref *ref)
#else /* CONFIG_MEM_ALLOC_PROFILING_DEBUG */ #else /* CONFIG_MEM_ALLOC_PROFILING_DEBUG */
static inline bool is_codetag_empty(union codetag_ref *ref) { return false; } static inline bool is_codetag_empty(union codetag_ref *ref) { return false; }
static inline void set_codetag_empty(union codetag_ref *ref) {}
#endif /* CONFIG_MEM_ALLOC_PROFILING_DEBUG */ #endif /* CONFIG_MEM_ALLOC_PROFILING_DEBUG */
......
...@@ -5,6 +5,7 @@ ...@@ -5,6 +5,7 @@
#include <linux/errno.h> #include <linux/errno.h>
#include <linux/mmdebug.h> #include <linux/mmdebug.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <linux/pgalloc_tag.h>
#include <linux/bug.h> #include <linux/bug.h>
#include <linux/list.h> #include <linux/list.h>
#include <linux/mmzone.h> #include <linux/mmzone.h>
...@@ -3134,6 +3135,14 @@ extern void reserve_bootmem_region(phys_addr_t start, ...@@ -3134,6 +3135,14 @@ extern void reserve_bootmem_region(phys_addr_t start,
/* Free the reserved page into the buddy system, so it gets managed. */ /* Free the reserved page into the buddy system, so it gets managed. */
static inline void free_reserved_page(struct page *page) static inline void free_reserved_page(struct page *page)
{ {
if (mem_alloc_profiling_enabled()) {
union codetag_ref *ref = get_page_tag_ref(page);
if (ref) {
set_codetag_empty(ref);
put_page_tag_ref(ref);
}
}
ClearPageReserved(page); ClearPageReserved(page);
init_page_count(page); init_page_count(page);
__free_page(page); __free_page(page);
......
...@@ -120,6 +120,8 @@ static inline void pgalloc_tag_sub_pages(struct alloc_tag *tag, unsigned int nr) ...@@ -120,6 +120,8 @@ static inline void pgalloc_tag_sub_pages(struct alloc_tag *tag, unsigned int nr)
#else /* CONFIG_MEM_ALLOC_PROFILING */ #else /* CONFIG_MEM_ALLOC_PROFILING */
static inline union codetag_ref *get_page_tag_ref(struct page *page) { return NULL; }
static inline void put_page_tag_ref(union codetag_ref *ref) {}
static inline void pgalloc_tag_add(struct page *page, struct task_struct *task, static inline void pgalloc_tag_add(struct page *page, struct task_struct *task,
unsigned int nr) {} unsigned int nr) {}
static inline void pgalloc_tag_sub(struct page *page, unsigned int nr) {} static inline void pgalloc_tag_sub(struct page *page, unsigned int nr) {}
......
...@@ -2567,7 +2567,6 @@ void __init set_dma_reserve(unsigned long new_dma_reserve) ...@@ -2567,7 +2567,6 @@ void __init set_dma_reserve(unsigned long new_dma_reserve)
void __init memblock_free_pages(struct page *page, unsigned long pfn, void __init memblock_free_pages(struct page *page, unsigned long pfn,
unsigned int order) unsigned int order)
{ {
if (IS_ENABLED(CONFIG_DEFERRED_STRUCT_PAGE_INIT)) { if (IS_ENABLED(CONFIG_DEFERRED_STRUCT_PAGE_INIT)) {
int nid = early_pfn_to_nid(pfn); int nid = early_pfn_to_nid(pfn);
...@@ -2579,6 +2578,17 @@ void __init memblock_free_pages(struct page *page, unsigned long pfn, ...@@ -2579,6 +2578,17 @@ void __init memblock_free_pages(struct page *page, unsigned long pfn,
/* KMSAN will take care of these pages. */ /* KMSAN will take care of these pages. */
return; return;
} }
/* pages were reserved and not allocated */
if (mem_alloc_profiling_enabled()) {
union codetag_ref *ref = get_page_tag_ref(page);
if (ref) {
set_codetag_empty(ref);
put_page_tag_ref(ref);
}
}
__free_pages_core(page, order); __free_pages_core(page, order);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment