| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * KMSAN hooks for kernel subsystems.
- *
- * These functions handle creation of KMSAN metadata for memory allocations.
- *
- * Copyright (C) 2018-2022 Google LLC
- * Author: Alexander Potapenko <glider@google.com>
- *
- */
- #include <linux/cacheflush.h>
- #include <linux/dma-direction.h>
- #include <linux/gfp.h>
- #include <linux/kmsan.h>
- #include <linux/mm.h>
- #include <linux/mm_types.h>
- #include <linux/scatterlist.h>
- #include <linux/slab.h>
- #include <linux/uaccess.h>
- #include <linux/usb.h>
- #include "../internal.h"
- #include "../slab.h"
- #include "kmsan.h"
- /*
- * Instrumented functions shouldn't be called under
- * kmsan_enter_runtime()/kmsan_leave_runtime(), because this will lead to
- * skipping effects of functions like memset() inside instrumented code.
- */
- void kmsan_task_create(struct task_struct *task)
- {
- kmsan_enter_runtime();
- kmsan_internal_task_create(task);
- kmsan_leave_runtime();
- }
- void kmsan_task_exit(struct task_struct *task)
- {
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- kmsan_disable_current();
- }
- void kmsan_slab_alloc(struct kmem_cache *s, void *object, gfp_t flags)
- {
- if (unlikely(object == NULL))
- return;
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- /*
- * There's a ctor or this is an RCU cache - do nothing. The memory
- * status hasn't changed since last use.
- */
- if (s->ctor || (s->flags & SLAB_TYPESAFE_BY_RCU))
- return;
- kmsan_enter_runtime();
- if (flags & __GFP_ZERO)
- kmsan_internal_unpoison_memory(object, s->object_size,
- KMSAN_POISON_CHECK);
- else
- kmsan_internal_poison_memory(object, s->object_size, flags,
- KMSAN_POISON_CHECK);
- kmsan_leave_runtime();
- }
- void kmsan_slab_free(struct kmem_cache *s, void *object)
- {
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- /* RCU slabs could be legally used after free within the RCU period */
- if (unlikely(s->flags & SLAB_TYPESAFE_BY_RCU))
- return;
- /*
- * If there's a constructor, freed memory must remain in the same state
- * until the next allocation. We cannot save its state to detect
- * use-after-free bugs, instead we just keep it unpoisoned.
- */
- if (s->ctor)
- return;
- kmsan_enter_runtime();
- kmsan_internal_poison_memory(object, s->object_size, GFP_KERNEL,
- KMSAN_POISON_CHECK | KMSAN_POISON_FREE);
- kmsan_leave_runtime();
- }
- void kmsan_kmalloc_large(const void *ptr, size_t size, gfp_t flags)
- {
- if (unlikely(ptr == NULL))
- return;
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- kmsan_enter_runtime();
- if (flags & __GFP_ZERO)
- kmsan_internal_unpoison_memory((void *)ptr, size,
- /*checked*/ true);
- else
- kmsan_internal_poison_memory((void *)ptr, size, flags,
- KMSAN_POISON_CHECK);
- kmsan_leave_runtime();
- }
- void kmsan_kfree_large(const void *ptr)
- {
- struct page *page;
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- kmsan_enter_runtime();
- page = virt_to_head_page((void *)ptr);
- KMSAN_WARN_ON(ptr != page_address(page));
- kmsan_internal_poison_memory((void *)ptr,
- page_size(page),
- GFP_KERNEL,
- KMSAN_POISON_CHECK | KMSAN_POISON_FREE);
- kmsan_leave_runtime();
- }
- static unsigned long vmalloc_shadow(unsigned long addr)
- {
- return (unsigned long)kmsan_get_metadata((void *)addr,
- KMSAN_META_SHADOW);
- }
- static unsigned long vmalloc_origin(unsigned long addr)
- {
- return (unsigned long)kmsan_get_metadata((void *)addr,
- KMSAN_META_ORIGIN);
- }
- void kmsan_vunmap_range_noflush(unsigned long start, unsigned long end)
- {
- __vunmap_range_noflush(vmalloc_shadow(start), vmalloc_shadow(end));
- __vunmap_range_noflush(vmalloc_origin(start), vmalloc_origin(end));
- flush_cache_vmap(vmalloc_shadow(start), vmalloc_shadow(end));
- flush_cache_vmap(vmalloc_origin(start), vmalloc_origin(end));
- }
- /*
- * This function creates new shadow/origin pages for the physical pages mapped
- * into the virtual memory. If those physical pages already had shadow/origin,
- * those are ignored.
- */
- int kmsan_ioremap_page_range(unsigned long start, unsigned long end,
- phys_addr_t phys_addr, pgprot_t prot,
- unsigned int page_shift)
- {
- gfp_t gfp_mask = GFP_KERNEL | __GFP_ZERO;
- struct page *shadow, *origin;
- unsigned long off = 0;
- int nr, err = 0, clean = 0, mapped;
- if (!kmsan_enabled || kmsan_in_runtime())
- return 0;
- nr = (end - start) / PAGE_SIZE;
- kmsan_enter_runtime();
- for (int i = 0; i < nr; i++, off += PAGE_SIZE, clean = i) {
- shadow = alloc_pages(gfp_mask, 1);
- origin = alloc_pages(gfp_mask, 1);
- if (!shadow || !origin) {
- err = -ENOMEM;
- goto ret;
- }
- mapped = __vmap_pages_range_noflush(
- vmalloc_shadow(start + off),
- vmalloc_shadow(start + off + PAGE_SIZE), prot, &shadow,
- PAGE_SHIFT);
- if (mapped) {
- err = mapped;
- goto ret;
- }
- shadow = NULL;
- mapped = __vmap_pages_range_noflush(
- vmalloc_origin(start + off),
- vmalloc_origin(start + off + PAGE_SIZE), prot, &origin,
- PAGE_SHIFT);
- if (mapped) {
- __vunmap_range_noflush(
- vmalloc_shadow(start + off),
- vmalloc_shadow(start + off + PAGE_SIZE));
- err = mapped;
- goto ret;
- }
- origin = NULL;
- }
- /* Page mapping loop finished normally, nothing to clean up. */
- clean = 0;
- ret:
- if (clean > 0) {
- /*
- * Something went wrong. Clean up shadow/origin pages allocated
- * on the last loop iteration, then delete mappings created
- * during the previous iterations.
- */
- if (shadow)
- __free_pages(shadow, 1);
- if (origin)
- __free_pages(origin, 1);
- __vunmap_range_noflush(
- vmalloc_shadow(start),
- vmalloc_shadow(start + clean * PAGE_SIZE));
- __vunmap_range_noflush(
- vmalloc_origin(start),
- vmalloc_origin(start + clean * PAGE_SIZE));
- }
- flush_cache_vmap(vmalloc_shadow(start), vmalloc_shadow(end));
- flush_cache_vmap(vmalloc_origin(start), vmalloc_origin(end));
- kmsan_leave_runtime();
- return err;
- }
- void kmsan_iounmap_page_range(unsigned long start, unsigned long end)
- {
- unsigned long v_shadow, v_origin;
- struct page *shadow, *origin;
- int nr;
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- nr = (end - start) / PAGE_SIZE;
- kmsan_enter_runtime();
- v_shadow = (unsigned long)vmalloc_shadow(start);
- v_origin = (unsigned long)vmalloc_origin(start);
- for (int i = 0; i < nr;
- i++, v_shadow += PAGE_SIZE, v_origin += PAGE_SIZE) {
- shadow = kmsan_vmalloc_to_page_or_null((void *)v_shadow);
- origin = kmsan_vmalloc_to_page_or_null((void *)v_origin);
- __vunmap_range_noflush(v_shadow, vmalloc_shadow(end));
- __vunmap_range_noflush(v_origin, vmalloc_origin(end));
- if (shadow)
- __free_pages(shadow, 1);
- if (origin)
- __free_pages(origin, 1);
- }
- flush_cache_vmap(vmalloc_shadow(start), vmalloc_shadow(end));
- flush_cache_vmap(vmalloc_origin(start), vmalloc_origin(end));
- kmsan_leave_runtime();
- }
- void kmsan_copy_to_user(void __user *to, const void *from, size_t to_copy,
- size_t left)
- {
- unsigned long ua_flags;
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- /*
- * At this point we've copied the memory already. It's hard to check it
- * before copying, as the size of actually copied buffer is unknown.
- */
- /* copy_to_user() may copy zero bytes. No need to check. */
- if (!to_copy)
- return;
- /* Or maybe copy_to_user() failed to copy anything. */
- if (to_copy <= left)
- return;
- ua_flags = user_access_save();
- if (!IS_ENABLED(CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE) ||
- (u64)to < TASK_SIZE) {
- /* This is a user memory access, check it. */
- kmsan_internal_check_memory((void *)from, to_copy - left, to,
- REASON_COPY_TO_USER);
- } else {
- /* Otherwise this is a kernel memory access. This happens when a
- * compat syscall passes an argument allocated on the kernel
- * stack to a real syscall.
- * Don't check anything, just copy the shadow of the copied
- * bytes.
- */
- kmsan_internal_memmove_metadata((void *)to, (void *)from,
- to_copy - left);
- }
- user_access_restore(ua_flags);
- }
- EXPORT_SYMBOL(kmsan_copy_to_user);
- void kmsan_memmove(void *to, const void *from, size_t size)
- {
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- kmsan_enter_runtime();
- kmsan_internal_memmove_metadata(to, (void *)from, size);
- kmsan_leave_runtime();
- }
- EXPORT_SYMBOL(kmsan_memmove);
- /* Helper function to check an URB. */
- void kmsan_handle_urb(const struct urb *urb, bool is_out)
- {
- if (!urb)
- return;
- if (is_out)
- kmsan_internal_check_memory(urb->transfer_buffer,
- urb->transfer_buffer_length,
- /*user_addr*/ NULL,
- REASON_SUBMIT_URB);
- else
- kmsan_internal_unpoison_memory(urb->transfer_buffer,
- urb->transfer_buffer_length,
- /*checked*/ false);
- }
- EXPORT_SYMBOL_GPL(kmsan_handle_urb);
- static void kmsan_handle_dma_page(const void *addr, size_t size,
- enum dma_data_direction dir)
- {
- switch (dir) {
- case DMA_BIDIRECTIONAL:
- kmsan_internal_check_memory((void *)addr, size,
- /*user_addr*/ NULL, REASON_ANY);
- kmsan_internal_unpoison_memory((void *)addr, size,
- /*checked*/ false);
- break;
- case DMA_TO_DEVICE:
- kmsan_internal_check_memory((void *)addr, size,
- /*user_addr*/ NULL, REASON_ANY);
- break;
- case DMA_FROM_DEVICE:
- kmsan_internal_unpoison_memory((void *)addr, size,
- /*checked*/ false);
- break;
- case DMA_NONE:
- break;
- }
- }
- /* Helper function to handle DMA data transfers. */
- void kmsan_handle_dma(struct page *page, size_t offset, size_t size,
- enum dma_data_direction dir)
- {
- u64 page_offset, to_go, addr;
- if (PageHighMem(page))
- return;
- addr = (u64)page_address(page) + offset;
- /*
- * The kernel may occasionally give us adjacent DMA pages not belonging
- * to the same allocation. Process them separately to avoid triggering
- * internal KMSAN checks.
- */
- while (size > 0) {
- page_offset = offset_in_page(addr);
- to_go = min(PAGE_SIZE - page_offset, (u64)size);
- kmsan_handle_dma_page((void *)addr, to_go, dir);
- addr += to_go;
- size -= to_go;
- }
- }
- void kmsan_handle_dma_sg(struct scatterlist *sg, int nents,
- enum dma_data_direction dir)
- {
- struct scatterlist *item;
- int i;
- for_each_sg(sg, item, nents, i)
- kmsan_handle_dma(sg_page(item), item->offset, item->length,
- dir);
- }
- /* Functions from kmsan-checks.h follow. */
- /*
- * To create an origin, kmsan_poison_memory() unwinds the stacks and stores it
- * into the stack depot. This may cause deadlocks if done from within KMSAN
- * runtime, therefore we bail out if kmsan_in_runtime().
- */
- void kmsan_poison_memory(const void *address, size_t size, gfp_t flags)
- {
- if (!kmsan_enabled || kmsan_in_runtime())
- return;
- kmsan_enter_runtime();
- /* The users may want to poison/unpoison random memory. */
- kmsan_internal_poison_memory((void *)address, size, flags,
- KMSAN_POISON_NOCHECK);
- kmsan_leave_runtime();
- }
- EXPORT_SYMBOL(kmsan_poison_memory);
- /*
- * Unlike kmsan_poison_memory(), this function can be used from within KMSAN
- * runtime, because it does not trigger allocations or call instrumented code.
- */
- void kmsan_unpoison_memory(const void *address, size_t size)
- {
- unsigned long ua_flags;
- if (!kmsan_enabled)
- return;
- ua_flags = user_access_save();
- /* The users may want to poison/unpoison random memory. */
- kmsan_internal_unpoison_memory((void *)address, size,
- KMSAN_POISON_NOCHECK);
- user_access_restore(ua_flags);
- }
- EXPORT_SYMBOL(kmsan_unpoison_memory);
- /*
- * Version of kmsan_unpoison_memory() called from IRQ entry functions.
- */
- void kmsan_unpoison_entry_regs(const struct pt_regs *regs)
- {
- kmsan_unpoison_memory((void *)regs, sizeof(*regs));
- }
- void kmsan_check_memory(const void *addr, size_t size)
- {
- if (!kmsan_enabled)
- return;
- return kmsan_internal_check_memory((void *)addr, size,
- /*user_addr*/ NULL, REASON_ANY);
- }
- EXPORT_SYMBOL(kmsan_check_memory);
- void kmsan_enable_current(void)
- {
- KMSAN_WARN_ON(current->kmsan_ctx.depth == 0);
- current->kmsan_ctx.depth--;
- }
- EXPORT_SYMBOL(kmsan_enable_current);
- void kmsan_disable_current(void)
- {
- current->kmsan_ctx.depth++;
- KMSAN_WARN_ON(current->kmsan_ctx.depth == 0);
- }
- EXPORT_SYMBOL(kmsan_disable_current);
|