Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757558Ab1DAOol (ORCPT ); Fri, 1 Apr 2011 10:44:41 -0400 Received: from e28smtp03.in.ibm.com ([122.248.162.3]:42126 "EHLO e28smtp03.in.ibm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1757279Ab1DAOoi (ORCPT ); Fri, 1 Apr 2011 10:44:38 -0400 From: Srikar Dronamraju To: Peter Zijlstra , Ingo Molnar Cc: Steven Rostedt , Srikar Dronamraju , Linux-mm , Arnaldo Carvalho de Melo , Linus Torvalds , Jonathan Corbet , Christoph Hellwig , Masami Hiramatsu , Thomas Gleixner , Ananth N Mavinakayanahalli , Oleg Nesterov , Andrew Morton , SystemTap , Jim Keniston , Roland McGrath , Andi Kleen , LKML Date: Fri, 01 Apr 2011 20:04:57 +0530 Message-Id: <20110401143457.15455.64839.sendpatchset@localhost6.localdomain6> In-Reply-To: <20110401143223.15455.19844.sendpatchset@localhost6.localdomain6> References: <20110401143223.15455.19844.sendpatchset@localhost6.localdomain6> Subject: [PATCH v3 2.6.39-rc1-tip 12/26] 12: uprobes: slot allocation for uprobes Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 11016 Lines: 391 Every task is allocated a fixed slot. When a probe is hit, the original instruction corresponding to the probe hit is copied to per-task fixed slot. Currently we allocate one page of slots for each mm. Bitmaps are used to know which slots are free. Each slot is made of 128 bytes so that its cache aligned. TODO: On massively threaded processes (or if a huge number of processes share the same mm), there is a possiblilty of running out of slots. One alternative could be to extend the slots as when slots are required. Signed-off-by: Srikar Dronamraju Signed-off-by: Jim Keniston --- include/linux/mm_types.h | 4 + include/linux/uprobes.h | 21 ++++ kernel/fork.c | 4 + kernel/uprobes.c | 238 ++++++++++++++++++++++++++++++++++++++++++++++ 4 files changed, 267 insertions(+), 0 deletions(-) diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h index c691096..ff4c72b 100644 --- a/include/linux/mm_types.h +++ b/include/linux/mm_types.h @@ -12,6 +12,9 @@ #include #include #include +#ifdef CONFIG_UPROBES +#include +#endif #include #include @@ -321,6 +324,7 @@ struct mm_struct { unsigned long uprobes_vaddr; struct list_head uprobes_list; /* protected by uprobes_mutex */ atomic_t uprobes_count; + struct uprobes_xol_area *uprobes_xol_area; #endif }; diff --git a/include/linux/uprobes.h b/include/linux/uprobes.h index 8da993c..10647be 100644 --- a/include/linux/uprobes.h +++ b/include/linux/uprobes.h @@ -101,6 +101,25 @@ struct uprobe_task { }; /* + * Every thread gets its own slot. Once it's assigned a slot, it + * keeps that slot until the thread exits. Only definite number + * of slots are allocated. + */ + +struct uprobes_xol_area { + spinlock_t slot_lock; /* protects bitmap and slot (de)allocation*/ + unsigned long *bitmap; /* 0 = free slot */ + struct page *page; + + /* + * We keep the vma's vm_start rather than a pointer to the vma + * itself. The probed process or a naughty kernel module could make + * the vma go away, and we must handle that reasonably gracefully. + */ + unsigned long vaddr; /* Page(s) of instruction slots */ +}; + +/* * Most architectures can use the default versions of @read_opcode(), * @set_bkpt(), @set_orig_insn(), and @is_bkpt_insn(); * @@ -139,6 +158,7 @@ extern void uprobe_free_utask(struct task_struct *tsk); struct vm_area_struct; extern int uprobe_mmap(struct vm_area_struct *vma); extern void uprobe_dup_mmap(struct mm_struct *old_mm, struct mm_struct *mm); +extern void uprobes_free_xol_area(struct mm_struct *mm); #else /* CONFIG_UPROBES is not defined */ static inline int register_uprobe(struct inode *inode, loff_t offset, struct uprobe_consumer *consumer) @@ -158,5 +178,6 @@ static inline int uprobe_mmap(struct vm_area_struct *vma) return 0; } static inline void uprobe_free_utask(struct task_struct *tsk) {} +static inline void uprobes_free_xol_area(struct mm_struct *mm) {} #endif /* CONFIG_UPROBES */ #endif /* _LINUX_UPROBES_H */ diff --git a/kernel/fork.c b/kernel/fork.c index e25c29e..7131096 100644 --- a/kernel/fork.c +++ b/kernel/fork.c @@ -558,6 +558,7 @@ void mmput(struct mm_struct *mm) might_sleep(); if (atomic_dec_and_test(&mm->mm_users)) { + uprobes_free_xol_area(mm); exit_aio(mm); ksm_exit(mm); khugepaged_exit(mm); /* must run before exit_mmap */ @@ -690,6 +691,9 @@ struct mm_struct *dup_mm(struct task_struct *tsk) #ifdef CONFIG_TRANSPARENT_HUGEPAGE mm->pmd_huge_pte = NULL; #endif +#ifdef CONFIG_UPROBES + mm->uprobes_xol_area = NULL; +#endif if (!mm_init(mm, tsk)) goto fail_nomem; diff --git a/kernel/uprobes.c b/kernel/uprobes.c index f9fb7c2..7663d18 100644 --- a/kernel/uprobes.c +++ b/kernel/uprobes.c @@ -31,12 +31,28 @@ #include #include #include /* needed for anon_vma_prepare */ +#include /* needed for PROT_EXEC, MAP_PRIVATE */ +#include /* needed for fput() */ +#define UINSNS_PER_PAGE (PAGE_SIZE/UPROBES_XOL_SLOT_BYTES) +#define MAX_UPROBES_XOL_SLOTS UINSNS_PER_PAGE + +/* + * valid_vma: Verify if the specified vma is an executable vma, + * but not an XOL vma. + * - Return 1 if the specified virtual address is in an + * executable vma, but not in an XOL vma. + */ static bool valid_vma(struct vm_area_struct *vma) { + struct uprobes_xol_area *area = vma->vm_mm->uprobes_xol_area; + if (!vma->vm_file) return false; + if (area && (area->vaddr == vma->vm_start)) + return false; + if ((vma->vm_flags & (VM_READ|VM_WRITE|VM_EXEC|VM_SHARED)) == (VM_READ|VM_EXEC)) return true; @@ -956,6 +972,224 @@ mmap_out: return ret; } +/* Slot allocation for XOL */ + +static int xol_add_vma(struct uprobes_xol_area *area) +{ + struct vm_area_struct *vma; + struct mm_struct *mm; + struct file *file; + unsigned long addr; + int ret = -ENOMEM; + + mm = get_task_mm(current); + if (!mm) + return -ESRCH; + + down_write(&mm->mmap_sem); + if (mm->uprobes_xol_area) { + ret = -EALREADY; + goto fail; + } + + /* + * Find the end of the top mapping and skip a page. + * If there is no space for PAGE_SIZE above + * that, mmap will ignore our address hint. + * + * We allocate a "fake" unlinked shmem file because + * anonymous memory might not be granted execute + * permission when the selinux security hooks have + * their way. + */ + vma = rb_entry(rb_last(&mm->mm_rb), struct vm_area_struct, vm_rb); + addr = vma->vm_end + PAGE_SIZE; + file = shmem_file_setup("uprobes/xol", PAGE_SIZE, VM_NORESERVE); + if (!file) { + printk(KERN_ERR "uprobes_xol failed to setup shmem_file " + "while allocating vma for pid/tgid %d/%d for " + "single-stepping out of line.\n", + current->pid, current->tgid); + goto fail; + } + addr = do_mmap_pgoff(file, addr, PAGE_SIZE, PROT_EXEC, MAP_PRIVATE, 0); + fput(file); + + if (addr & ~PAGE_MASK) { + printk(KERN_ERR "uprobes_xol failed to allocate a vma for " + "pid/tgid %d/%d for single-stepping out of " + "line.\n", current->pid, current->tgid); + goto fail; + } + vma = find_vma(mm, addr); + + /* Don't expand vma on mremap(). */ + vma->vm_flags |= VM_DONTEXPAND | VM_DONTCOPY; + area->vaddr = vma->vm_start; + if (get_user_pages(current, mm, area->vaddr, 1, 1, 1, &area->page, + &vma) > 0) + ret = 0; + +fail: + up_write(&mm->mmap_sem); + mmput(mm); + return ret; +} + +/* + * xol_alloc_area - Allocate process's uprobes_xol_area. + * This area will be used for storing instructions for execution out of + * line. + * + * Returns the allocated area or NULL. + */ +static struct uprobes_xol_area *xol_alloc_area(void) +{ + struct uprobes_xol_area *area = NULL; + + area = kzalloc(sizeof(*area), GFP_KERNEL); + if (unlikely(!area)) + return NULL; + + area->bitmap = kzalloc(BITS_TO_LONGS(UINSNS_PER_PAGE) * sizeof(long), + GFP_KERNEL); + + if (!area->bitmap) + goto fail; + + spin_lock_init(&area->slot_lock); + if (!xol_add_vma(area) && !current->mm->uprobes_xol_area) { + task_lock(current); + if (!current->mm->uprobes_xol_area) { + current->mm->uprobes_xol_area = area; + task_unlock(current); + return area; + } + task_unlock(current); + } + +fail: + kfree(area->bitmap); + kfree(area); + return current->mm->uprobes_xol_area; +} + +/* + * uprobes_free_xol_area - Free the area allocated for slots. + */ +void uprobes_free_xol_area(struct mm_struct *mm) +{ + struct uprobes_xol_area *area = mm->uprobes_xol_area; + + if (!area) + return; + + put_page(area->page); + kfree(area->bitmap); + kfree(area); +} + +/* + * Find a slot + * - searching in existing vmas for a free slot. + * - If no free slot in existing vmas, return 0; + * + * Called when holding uprobes_xol_area->slot_lock + */ +static unsigned long xol_take_insn_slot(struct uprobes_xol_area *area) +{ + unsigned long slot_addr; + int slot_nr; + + slot_nr = find_first_zero_bit(area->bitmap, UINSNS_PER_PAGE); + if (slot_nr < UINSNS_PER_PAGE) { + __set_bit(slot_nr, area->bitmap); + slot_addr = area->vaddr + + (slot_nr * UPROBES_XOL_SLOT_BYTES); + return slot_addr; + } + + return 0; +} + +/* + * xol_get_insn_slot - If was not allocated a slot, then + * allocate a slot. + * Returns the allocated slot address or 0. + */ +static unsigned long xol_get_insn_slot(struct uprobe *uprobe, + unsigned long slot_addr) +{ + struct uprobes_xol_area *area = current->mm->uprobes_xol_area; + unsigned long flags, xol_vaddr = current->utask->xol_vaddr; + void *vaddr; + + if (!current->utask->xol_vaddr || !area) { + if (!area) + area = xol_alloc_area(); + + if (!area) + return 0; + + spin_lock_irqsave(&area->slot_lock, flags); + xol_vaddr = xol_take_insn_slot(area); + spin_unlock_irqrestore(&area->slot_lock, flags); + current->utask->xol_vaddr = xol_vaddr; + } + + /* + * Initialize the slot if xol_vaddr points to valid + * instruction slot. + */ + if (unlikely(!xol_vaddr)) + return 0; + + current->utask->vaddr = slot_addr; + vaddr = kmap_atomic(area->page, KM_USER0); + xol_vaddr &= ~PAGE_MASK; + memcpy(vaddr + xol_vaddr, uprobe->insn, MAX_UINSN_BYTES); + kunmap_atomic(vaddr, KM_USER0); + return current->utask->xol_vaddr; +} + +/* + * xol_free_insn_slot - If slot was earlier allocated by + * @xol_get_insn_slot(), make the slot available for + * subsequent requests. + */ +static void xol_free_insn_slot(struct task_struct *tsk, unsigned long slot_addr) +{ + struct uprobes_xol_area *area; + unsigned long vma_end; + + if (!tsk->mm || !tsk->mm->uprobes_xol_area) + return; + + area = tsk->mm->uprobes_xol_area; + + if (unlikely(!slot_addr || IS_ERR_VALUE(slot_addr))) + return; + + vma_end = area->vaddr + PAGE_SIZE; + if (area->vaddr <= slot_addr && slot_addr < vma_end) { + int slot_nr; + unsigned long offset = slot_addr - area->vaddr; + unsigned long flags; + + BUG_ON(offset % UPROBES_XOL_SLOT_BYTES); + + slot_nr = offset / UPROBES_XOL_SLOT_BYTES; + BUG_ON(slot_nr >= UINSNS_PER_PAGE); + + spin_lock_irqsave(&area->slot_lock, flags); + __clear_bit(slot_nr, area->bitmap); + spin_unlock_irqrestore(&area->slot_lock, flags); + return; + } + printk(KERN_ERR "%s: no XOL vma for slot address %#lx\n", + __func__, slot_addr); +} + /* * Called with no locks held. * Called in context of a exiting or a exec-ing thread. @@ -963,14 +1197,18 @@ mmap_out: void uprobe_free_utask(struct task_struct *tsk) { struct uprobe_task *utask = tsk->utask; + unsigned long xol_vaddr; if (!utask) return; + xol_vaddr = utask->xol_vaddr; if (utask->active_uprobe) put_uprobe(utask->active_uprobe); + kfree(utask); tsk->utask = NULL; + xol_free_insn_slot(tsk, xol_vaddr); } /* -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/