mm: page_alloc: add kasan hooks on alloc and free paths
authorAndrey Ryabinin <a.ryabinin@samsung.com>
Fri, 13 Feb 2015 22:39:28 +0000 (14:39 -0800)
committerLinus Torvalds <torvalds@linux-foundation.org>
Sat, 14 Feb 2015 05:21:41 +0000 (21:21 -0800)
Add kernel address sanitizer hooks to mark allocated page's addresses as
accessible in corresponding shadow region.  Mark freed pages as
inaccessible.

Signed-off-by: Andrey Ryabinin <a.ryabinin@samsung.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Konstantin Serebryany <kcc@google.com>
Cc: Dmitry Chernenkov <dmitryc@google.com>
Signed-off-by: Andrey Konovalov <adech.fo@gmail.com>
Cc: Yuri Gribov <tetra2005@gmail.com>
Cc: Konstantin Khlebnikov <koct9i@gmail.com>
Cc: Sasha Levin <sasha.levin@oracle.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Dave Hansen <dave.hansen@intel.com>
Cc: Andi Kleen <andi@firstfloor.org>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
include/linux/kasan.h
mm/compaction.c
mm/kasan/kasan.c
mm/kasan/kasan.h
mm/kasan/report.c
mm/page_alloc.c

index 9102fda60deff29d14a6577504ce5a641f42e1fd..f00c15c412354e523a477b4e0e37bbf68eb5de8f 100644 (file)
@@ -34,6 +34,9 @@ static inline void kasan_disable_current(void)
 
 void kasan_unpoison_shadow(const void *address, size_t size);
 
+void kasan_alloc_pages(struct page *page, unsigned int order);
+void kasan_free_pages(struct page *page, unsigned int order);
+
 #else /* CONFIG_KASAN */
 
 static inline void kasan_unpoison_shadow(const void *address, size_t size) {}
@@ -41,6 +44,9 @@ static inline void kasan_unpoison_shadow(const void *address, size_t size) {}
 static inline void kasan_enable_current(void) {}
 static inline void kasan_disable_current(void) {}
 
+static inline void kasan_alloc_pages(struct page *page, unsigned int order) {}
+static inline void kasan_free_pages(struct page *page, unsigned int order) {}
+
 #endif /* CONFIG_KASAN */
 
 #endif /* LINUX_KASAN_H */
index d50d6de6f1b68a7b7c785a02bec080c00ac5f7dd..8c0d9459b54a02042dd2caf9489566188d9ea908 100644 (file)
@@ -16,6 +16,7 @@
 #include <linux/sysfs.h>
 #include <linux/balloon_compaction.h>
 #include <linux/page-isolation.h>
+#include <linux/kasan.h>
 #include "internal.h"
 
 #ifdef CONFIG_COMPACTION
@@ -72,6 +73,7 @@ static void map_pages(struct list_head *list)
        list_for_each_entry(page, list, lru) {
                arch_alloc_page(page, 0);
                kernel_map_pages(page, 1, 1);
+               kasan_alloc_pages(page, 0);
        }
 }
 
index def81104772f0f198b8a84e4d2c0ad699c1c2099..b516eb8632b9e1cb67f3ae7b12b5c74aaa38ffb9 100644 (file)
@@ -254,6 +254,20 @@ static __always_inline void check_memory_region(unsigned long addr,
        kasan_report(addr, size, write, _RET_IP_);
 }
 
+void kasan_alloc_pages(struct page *page, unsigned int order)
+{
+       if (likely(!PageHighMem(page)))
+               kasan_unpoison_shadow(page_address(page), PAGE_SIZE << order);
+}
+
+void kasan_free_pages(struct page *page, unsigned int order)
+{
+       if (likely(!PageHighMem(page)))
+               kasan_poison_shadow(page_address(page),
+                               PAGE_SIZE << order,
+                               KASAN_FREE_PAGE);
+}
+
 #define DEFINE_ASAN_LOAD_STORE(size)                           \
        void __asan_load##size(unsigned long addr)              \
        {                                                       \
index 648b9c006f3f63ea278fc62b900b11ad4b0dc55c..d3c90d5dd97a274f8a1f615e28b62fd78752caae 100644 (file)
@@ -6,6 +6,8 @@
 #define KASAN_SHADOW_SCALE_SIZE (1UL << KASAN_SHADOW_SCALE_SHIFT)
 #define KASAN_SHADOW_MASK       (KASAN_SHADOW_SCALE_SIZE - 1)
 
+#define KASAN_FREE_PAGE         0xFF  /* page was freed */
+
 struct kasan_access_info {
        const void *access_addr;
        const void *first_bad_addr;
index 5835d69563f556751435e28dc6b63478653a732d..fab8e7882ff104bcc13d3ae7b40af4d4284efbdd 100644 (file)
@@ -54,6 +54,9 @@ static void print_error_description(struct kasan_access_info *info)
        shadow_val = *(u8 *)kasan_mem_to_shadow(info->first_bad_addr);
 
        switch (shadow_val) {
+       case KASAN_FREE_PAGE:
+               bug_type = "use after free";
+               break;
        case 0 ... KASAN_SHADOW_SCALE_SIZE - 1:
                bug_type = "out of bounds access";
                break;
@@ -69,6 +72,14 @@ static void print_error_description(struct kasan_access_info *info)
 
 static void print_address_description(struct kasan_access_info *info)
 {
+       const void *addr = info->access_addr;
+
+       if ((addr >= (void *)PAGE_OFFSET) &&
+               (addr < high_memory)) {
+               struct page *page = virt_to_head_page(addr);
+               dump_page(page, "kasan: bad access detected");
+       }
+
        dump_stack();
 }
 
index cb4758263f6b8a573251ae9323aad2160280ff34..a47f0b229a1aca202b15195c88ab93d52e65064f 100644 (file)
@@ -25,6 +25,7 @@
 #include <linux/compiler.h>
 #include <linux/kernel.h>
 #include <linux/kmemcheck.h>
+#include <linux/kasan.h>
 #include <linux/module.h>
 #include <linux/suspend.h>
 #include <linux/pagevec.h>
@@ -787,6 +788,7 @@ static bool free_pages_prepare(struct page *page, unsigned int order)
 
        trace_mm_page_free(page, order);
        kmemcheck_free_shadow(page, order);
+       kasan_free_pages(page, order);
 
        if (PageAnon(page))
                page->mapping = NULL;
@@ -970,6 +972,7 @@ static int prep_new_page(struct page *page, unsigned int order, gfp_t gfp_flags,
 
        arch_alloc_page(page, order);
        kernel_map_pages(page, 1 << order, 1);
+       kasan_alloc_pages(page, order);
 
        if (gfp_flags & __GFP_ZERO)
                prep_zero_page(page, order, gfp_flags);