Commit c80a0366 authored by Andrey Konovalov's avatar Andrey Konovalov Committed by Linus Torvalds

kasan: inline HW_TAGS helper functions

Mark all static functions in common.c and kasan.h that are used for
hardware tag-based KASAN as inline to avoid unnecessary function calls.

Link: https://lkml.kernel.org/r/2c94a2af0657f2b95b9337232339ff5ffa643ab5.1612546384.git.andreyknvl@google.comSigned-off-by: default avatarAndrey Konovalov <andreyknvl@google.com>
Reviewed-by: default avatarMarco Elver <elver@google.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Andrey Ryabinin <aryabinin@virtuozzo.com>
Cc: Branislav Rankov <Branislav.Rankov@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Evgenii Stepanov <eugenis@google.com>
Cc: Kevin Brodsky <kevin.brodsky@arm.com>
Cc: Peter Collingbourne <pcc@google.com>
Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 2cb34276
...@@ -279,7 +279,8 @@ void __kasan_poison_object_data(struct kmem_cache *cache, void *object) ...@@ -279,7 +279,8 @@ void __kasan_poison_object_data(struct kmem_cache *cache, void *object)
* based on objects indexes, so that objects that are next to each other * based on objects indexes, so that objects that are next to each other
* get different tags. * get different tags.
*/ */
static u8 assign_tag(struct kmem_cache *cache, const void *object, bool init) static inline u8 assign_tag(struct kmem_cache *cache,
const void *object, bool init)
{ {
if (IS_ENABLED(CONFIG_KASAN_GENERIC)) if (IS_ENABLED(CONFIG_KASAN_GENERIC))
return 0xff; return 0xff;
...@@ -321,8 +322,8 @@ void * __must_check __kasan_init_slab_obj(struct kmem_cache *cache, ...@@ -321,8 +322,8 @@ void * __must_check __kasan_init_slab_obj(struct kmem_cache *cache,
return (void *)object; return (void *)object;
} }
static bool ____kasan_slab_free(struct kmem_cache *cache, void *object, static inline bool ____kasan_slab_free(struct kmem_cache *cache,
unsigned long ip, bool quarantine) void *object, unsigned long ip, bool quarantine)
{ {
u8 tag; u8 tag;
void *tagged_object; void *tagged_object;
...@@ -366,7 +367,7 @@ bool __kasan_slab_free(struct kmem_cache *cache, void *object, unsigned long ip) ...@@ -366,7 +367,7 @@ bool __kasan_slab_free(struct kmem_cache *cache, void *object, unsigned long ip)
return ____kasan_slab_free(cache, object, ip, true); return ____kasan_slab_free(cache, object, ip, true);
} }
static bool ____kasan_kfree_large(void *ptr, unsigned long ip) static inline bool ____kasan_kfree_large(void *ptr, unsigned long ip)
{ {
if (ptr != page_address(virt_to_head_page(ptr))) { if (ptr != page_address(virt_to_head_page(ptr))) {
kasan_report_invalid_free(ptr, ip); kasan_report_invalid_free(ptr, ip);
...@@ -461,8 +462,8 @@ void * __must_check __kasan_slab_alloc(struct kmem_cache *cache, ...@@ -461,8 +462,8 @@ void * __must_check __kasan_slab_alloc(struct kmem_cache *cache,
return tagged_object; return tagged_object;
} }
static void *____kasan_kmalloc(struct kmem_cache *cache, const void *object, static inline void *____kasan_kmalloc(struct kmem_cache *cache,
size_t size, gfp_t flags) const void *object, size_t size, gfp_t flags)
{ {
unsigned long redzone_start; unsigned long redzone_start;
unsigned long redzone_end; unsigned long redzone_end;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment