From patchwork Fri Oct 29 12:27:44 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jianhua Liu X-Patchwork-Id: 12592623 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1ECE7C433F5 for ; Fri, 29 Oct 2021 12:30:38 +0000 (UTC) Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id D8D2261100 for ; Fri, 29 Oct 2021 12:30:37 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 mail.kernel.org D8D2261100 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=gmail.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20210309; h=Sender: Content-Transfer-Encoding:Content-Type:MIME-Version:List-Subscribe:List-Help: List-Post:List-Archive:List-Unsubscribe:List-Id:Message-Id:Date:Subject:Cc:To :From:Reply-To:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:In-Reply-To:References: List-Owner; bh=KuTyrhtZAEbo8e98/QhWra4HezoC2UFVYSzDdk0mlB0=; b=B2On+RhIuM4Jyl 7hFxD8rRvR0W8zOe6jQSETRyRC8gufnj7dCQXrTgBd8RIDq42Ur40rquuBaNacoiCKOAEjpWiYeRq HuuMz8B89MXGUdjmo3XfzuD7YmG/1qOlq2SKn26N5268WZ+UeIE8VHvr5tXI8p4cVXM48Zc9HoEjo 0R18O9yV2T6a7e5utP0qkij+JQp/25yXJGHv+5nNUOeuJzFAa7ikqC6PPxSxz8P7yYuVpGVvk2RSo YW0nTSqMuY1Wex/93q1LVmzKl4InUw59yqbh8slf+NVEY+xMKTgLu7NB/GOyUEWWNjbkC7N0XetN6 XKr7Zt9G+zWErcJjkG2w==; Received: from localhost ([::1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.94.2 #2 (Red Hat Linux)) id 1mgR0K-00AvGj-Im; Fri, 29 Oct 2021 12:29:04 +0000 Received: from mail-pf1-x436.google.com ([2607:f8b0:4864:20::436]) by bombadil.infradead.org with esmtps (Exim 4.94.2 #2 (Red Hat Linux)) id 1mgR0F-00AvG3-MX for linux-arm-kernel@lists.infradead.org; Fri, 29 Oct 2021 12:29:01 +0000 Received: by mail-pf1-x436.google.com with SMTP id p40so4250446pfh.8 for ; Fri, 29 Oct 2021 05:28:58 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id; bh=myqgbJHnLuKkUjjJk39/uy49V+NFfspOW/x3SYJ4yLc=; b=pyvB3TNRXAWDjKVqjfsL3BjqlDVBRtSLwERMApZT/6hZdyNGbh3lZ/SkSgkX2Oz07e 2JsYBiCzXFyx7eML+Aa2IxEttXD8xmA2OHEY5LNsx9AJLHnuSEF64THNbNPWCyWuzZkT FgSviJNX5ux6el0ExjUtkJFdT4S1HaCseT2Jnzni/y1BR6kFfzSAKnihgw345R8TXJaY 6FkNk76lNseP/0z+Xz663o2Ti9kDa+IeZlvnl6RHXDp3LHWVrN51oR5GGivEFsf0yizz 9/j0NWTUKNvwY9wanEnjOfY2uh6AgR96S9gRXnV4Tz5kXJebP5ChIQtM8oDPfIp25QtV sFkA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id; bh=myqgbJHnLuKkUjjJk39/uy49V+NFfspOW/x3SYJ4yLc=; b=Yj4YDNkFM/qb7nlQy5hzbGYDBZ5v/vK85aN02y0qKmBzFGnwHUxGUnd3jWQmGXI9pQ /43M/cnedpgGvCd4s/Tw8pfzS/3OyxC1u7oCLj0nFf46tygAUHIxSNHdlh6OvVOh+9ba 9TmECwOhO/U0bXoMAY0W9XQkYQQOV9/99mrJHIeA+EYJQgdreCUatMyHd/kZWLeO7wyl XnoOp9dT6yLCNgzwXU+1T7+/n7/o2P/wy37xu4R9zV6ofWF3DGDTGz8MNXks9TVOvcgO 7iQDDQQqsZD2d3DzaunBAsBHMNFgbQFSx/TqBUCLZyt3L8Yt2xKYi+eQB/16W8BC1E6x tt5w== X-Gm-Message-State: AOAM533f16CLbE365UYIwEkOACsbtbfwD+sVsuWrKNLAiroqcWr9cHtD q6wLZxtN6fPknaY91oHmoh0= X-Google-Smtp-Source: ABdhPJwL02CjaTmOPgRYlrjMDpvK59xPPBoRO+yZfhh/m81WM7dl0RK/wPZNSyF9iLi7TRqI8tGa/g== X-Received: by 2002:a05:6a00:168c:b0:44c:dc5a:7ee0 with SMTP id k12-20020a056a00168c00b0044cdc5a7ee0mr10851141pfc.40.1635510537764; Fri, 29 Oct 2021 05:28:57 -0700 (PDT) Received: from bj04871pcu.spreadtrum.com ([117.18.48.102]) by smtp.gmail.com with ESMTPSA id g8sm7033143pfv.123.2021.10.29.05.28.54 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Fri, 29 Oct 2021 05:28:56 -0700 (PDT) From: Janet Liu To: catalin.marinas@arm.com, will@kernel.org Cc: linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, orson.zhai@unisoc.com Subject: [PATCH v1 1/2] arm64: kprobes: implement optprobes Date: Fri, 29 Oct 2021 20:27:44 +0800 Message-Id: <1635510465-20469-1-git-send-email-jianhua.ljh@gmail.com> X-Mailer: git-send-email 1.9.1 X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20211029_052859_787043_2B3F9A74 X-CRM114-Status: GOOD ( 27.76 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.34 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , MIME-Version: 1.0 Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org From: Janet Liu Limitations: -only support 17(4096/240) kprobes 240 is from optprobe_template_end - optprobe_template_entry 4096 is from optinsn_slot reserve space size -only support steppable insn to be probed This patch replaced the probed instruction with a 'b' instruction, unconditionally branch to a buffer. The buffer contains instructions to create an pt_regs on stack, and then calls optimized_callback() which call the pre_handle(). After the executing kprobe handler, run the replaced instrunction, and branch to PC that probed instruction. The range of 'b' instruction is +/-128MB, alloc page from buddy is probable out of this +/-128MB range, so the buffer is allocated from a reserved area. For simple, only 4K is reserved. Futher patch can make optimization. Signed-off-by: Janet Liu Reported-by: kernel test robot --- arch/arm64/Kconfig | 1 + arch/arm64/include/asm/probes.h | 23 +++ arch/arm64/kernel/probes/Makefile | 1 + arch/arm64/kernel/probes/base_regs.h | 76 ++++++++ arch/arm64/kernel/probes/kprobes_trampoline.S | 55 +----- arch/arm64/kernel/probes/opt.c | 247 ++++++++++++++++++++++++++ arch/arm64/kernel/probes/opt_head.S | 40 +++++ 7 files changed, 389 insertions(+), 54 deletions(-) create mode 100644 arch/arm64/kernel/probes/base_regs.h create mode 100644 arch/arm64/kernel/probes/opt.c create mode 100644 arch/arm64/kernel/probes/opt_head.S diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig index 5c7ae4c..ab07251 100644 --- a/arch/arm64/Kconfig +++ b/arch/arm64/Kconfig @@ -199,6 +199,7 @@ config ARM64 select HAVE_STACKPROTECTOR select HAVE_SYSCALL_TRACEPOINTS select HAVE_KPROBES + select HAVE_OPTPROBES select HAVE_KRETPROBES select HAVE_GENERIC_VDSO select IOMMU_DMA if IOMMU_SUPPORT diff --git a/arch/arm64/include/asm/probes.h b/arch/arm64/include/asm/probes.h index 0069467..3114881 100644 --- a/arch/arm64/include/asm/probes.h +++ b/arch/arm64/include/asm/probes.h @@ -25,6 +25,29 @@ struct arch_probe_insn { struct arch_specific_insn { struct arch_probe_insn api; }; + +/* optinsn template addresses */ +extern __visible kprobe_opcode_t optinsn_slot; +extern __visible kprobe_opcode_t optprobe_template_entry; +extern __visible kprobe_opcode_t optprobe_template_val; +extern __visible kprobe_opcode_t optprobe_template_call; +extern __visible kprobe_opcode_t optprobe_template_end; +extern __visible kprobe_opcode_t optprobe_template_restore_orig_insn; +extern __visible kprobe_opcode_t optprobe_template_restore_end; + +#define MAX_OPTIMIZED_LENGTH 4 +#define MAX_OPTINSN_SIZE \ + ((kprobe_opcode_t *)&optprobe_template_end - \ + (kprobe_opcode_t *)&optprobe_template_entry) + + +struct arch_optimized_insn { + /* copy of the original instructions */ + kprobe_opcode_t copied_insn[AARCH64_INSN_SIZE]; + /* detour code buffer */ + kprobe_opcode_t *insn; +}; + #endif #endif diff --git a/arch/arm64/kernel/probes/Makefile b/arch/arm64/kernel/probes/Makefile index 8e4be92..c77c92a 100644 --- a/arch/arm64/kernel/probes/Makefile +++ b/arch/arm64/kernel/probes/Makefile @@ -2,5 +2,6 @@ obj-$(CONFIG_KPROBES) += kprobes.o decode-insn.o \ kprobes_trampoline.o \ simulate-insn.o +obj-$(CONFIG_OPTPROBES) += opt.o opt_head.o obj-$(CONFIG_UPROBES) += uprobes.o decode-insn.o \ simulate-insn.o diff --git a/arch/arm64/kernel/probes/base_regs.h b/arch/arm64/kernel/probes/base_regs.h new file mode 100644 index 0000000..45565dc --- /dev/null +++ b/arch/arm64/kernel/probes/base_regs.h @@ -0,0 +1,76 @@ +/* SPDX-License-Identifier: GPL-2.0-only */ +// +// Copyright (C) 2021, Unisoc Inc. +// Author: Janet Liu + +#include + + .macro save_all_base_regs + + sub sp, sp, #(PT_REGS_SIZE + 16) + + stp x0, x1, [sp, #S_X0] + stp x2, x3, [sp, #S_X2] + stp x4, x5, [sp, #S_X4] + stp x6, x7, [sp, #S_X6] + stp x8, x9, [sp, #S_X8] + stp x10, x11, [sp, #S_X10] + stp x12, x13, [sp, #S_X12] + stp x14, x15, [sp, #S_X14] + stp x16, x17, [sp, #S_X16] + stp x18, x19, [sp, #S_X18] + stp x20, x21, [sp, #S_X20] + stp x22, x23, [sp, #S_X22] + stp x24, x25, [sp, #S_X24] + stp x26, x27, [sp, #S_X26] + stp x28, x29, [sp, #S_X28] + add x0, sp, #(PT_REGS_SIZE + 16) + stp lr, x0, [sp, #S_LR] + + stp x29, x30, [sp, #PT_REGS_SIZE] + add x29, sp, #PT_REGS_SIZE + stp x29, x30, [sp, #S_STACKFRAME] + add x29, sp, #S_STACKFRAME + + /* + * Construct a useful saved PSTATE + */ + mrs x0, nzcv + mrs x1, daif + orr x0, x0, x1 + mrs x1, CurrentEL + orr x0, x0, x1 + mrs x1, SPSel + orr x0, x0, x1 + stp xzr, x0, [sp, #S_PC] + .endm + + .macro restore_all_base_regs trampoline = 0 + .if trampoline == 0 + ldr x0, [sp, #S_PSTATE] + and x0, x0, #(PSR_N_BIT | PSR_Z_BIT | PSR_C_BIT | PSR_V_BIT) + msr nzcv, x0 + .endif + + ldp x0, x1, [sp, #S_X0] + ldp x2, x3, [sp, #S_X2] + ldp x4, x5, [sp, #S_X4] + ldp x6, x7, [sp, #S_X6] + ldp x8, x9, [sp, #S_X8] + ldp x10, x11, [sp, #S_X10] + ldp x12, x13, [sp, #S_X12] + ldp x14, x15, [sp, #S_X14] + ldp x16, x17, [sp, #S_X16] + ldp x18, x19, [sp, #S_X18] + ldp x20, x21, [sp, #S_X20] + ldp x22, x23, [sp, #S_X22] + ldp x24, x25, [sp, #S_X24] + ldp x26, x27, [sp, #S_X26] + ldp x28, x29, [sp, #S_X28] + + .if \trampoline == 1 + ldr lr, [sp, #S_LR] + .endif + + add sp, sp, #(PT_REGS_SIZE + 16) + .endm diff --git a/arch/arm64/kernel/probes/kprobes_trampoline.S b/arch/arm64/kernel/probes/kprobes_trampoline.S index 288a84e2..cdc874f 100644 --- a/arch/arm64/kernel/probes/kprobes_trampoline.S +++ b/arch/arm64/kernel/probes/kprobes_trampoline.S @@ -6,63 +6,11 @@ #include #include #include +#include "base_regs.h" .text - .macro save_all_base_regs - stp x0, x1, [sp, #S_X0] - stp x2, x3, [sp, #S_X2] - stp x4, x5, [sp, #S_X4] - stp x6, x7, [sp, #S_X6] - stp x8, x9, [sp, #S_X8] - stp x10, x11, [sp, #S_X10] - stp x12, x13, [sp, #S_X12] - stp x14, x15, [sp, #S_X14] - stp x16, x17, [sp, #S_X16] - stp x18, x19, [sp, #S_X18] - stp x20, x21, [sp, #S_X20] - stp x22, x23, [sp, #S_X22] - stp x24, x25, [sp, #S_X24] - stp x26, x27, [sp, #S_X26] - stp x28, x29, [sp, #S_X28] - add x0, sp, #PT_REGS_SIZE - stp lr, x0, [sp, #S_LR] - /* - * Construct a useful saved PSTATE - */ - mrs x0, nzcv - mrs x1, daif - orr x0, x0, x1 - mrs x1, CurrentEL - orr x0, x0, x1 - mrs x1, SPSel - orr x0, x0, x1 - stp xzr, x0, [sp, #S_PC] - .endm - - .macro restore_all_base_regs - ldr x0, [sp, #S_PSTATE] - and x0, x0, #(PSR_N_BIT | PSR_Z_BIT | PSR_C_BIT | PSR_V_BIT) - msr nzcv, x0 - ldp x0, x1, [sp, #S_X0] - ldp x2, x3, [sp, #S_X2] - ldp x4, x5, [sp, #S_X4] - ldp x6, x7, [sp, #S_X6] - ldp x8, x9, [sp, #S_X8] - ldp x10, x11, [sp, #S_X10] - ldp x12, x13, [sp, #S_X12] - ldp x14, x15, [sp, #S_X14] - ldp x16, x17, [sp, #S_X16] - ldp x18, x19, [sp, #S_X18] - ldp x20, x21, [sp, #S_X20] - ldp x22, x23, [sp, #S_X22] - ldp x24, x25, [sp, #S_X24] - ldp x26, x27, [sp, #S_X26] - ldp x28, x29, [sp, #S_X28] - .endm - SYM_CODE_START(kretprobe_trampoline) - sub sp, sp, #PT_REGS_SIZE save_all_base_regs @@ -76,7 +24,6 @@ SYM_CODE_START(kretprobe_trampoline) restore_all_base_regs - add sp, sp, #PT_REGS_SIZE ret SYM_CODE_END(kretprobe_trampoline) diff --git a/arch/arm64/kernel/probes/opt.c b/arch/arm64/kernel/probes/opt.c new file mode 100644 index 0000000..b1f8f0d --- /dev/null +++ b/arch/arm64/kernel/probes/opt.c @@ -0,0 +1,247 @@ +// SPDX-License-Identifier: GPL-2.0-only +// +// Kernel Probes Jump Optimization (Optprobes) +// +// Copyright (C) 2021, Unisoc Inc. +// Author: Janet Liu +#include +#include +#include +#include +#include +#include +#include +#include + +#define TMPL_VAL_IDX \ + ((kprobe_opcode_t *)&optprobe_template_val - (kprobe_opcode_t *)&optprobe_template_entry) +#define TMPL_CALL_IDX \ + ((kprobe_opcode_t *)&optprobe_template_call - (kprobe_opcode_t *)&optprobe_template_entry) +#define TMPL_END_IDX \ + ((kprobe_opcode_t *)&optprobe_template_end - (kprobe_opcode_t *)&optprobe_template_entry) +#define TMPL_RESTORE_ORIGN_INSN \ + ((kprobe_opcode_t *)&optprobe_template_restore_orig_insn \ + - (kprobe_opcode_t *)&optprobe_template_entry) +#define TMPL_RESTORE_END \ + ((kprobe_opcode_t *)&optprobe_template_restore_end \ + - (kprobe_opcode_t *)&optprobe_template_entry) + + +static bool optinsn_page_in_use; + +void *alloc_optinsn_page(void) +{ + if (optinsn_page_in_use) + return NULL; + optinsn_page_in_use = true; + return &optinsn_slot; +} + +void free_optinsn_page(void *page __maybe_unused) +{ + optinsn_page_in_use = false; +} + +int arch_prepared_optinsn(struct arch_optimized_insn *optinsn) +{ + return optinsn->insn != NULL; +} + +/* + * In ARM64 ISA, kprobe opt always replace one instruction (4 bytes + * aligned and 4 bytes long). It is impossible to encounter another + * kprobe in the address range. So always return 0. + */ +int arch_check_optimized_kprobe(struct optimized_kprobe *op) +{ + return 0; +} + +/* only optimize steppable insn */ +static int can_optimize(struct kprobe *kp) +{ + if (!kp->ainsn.api.insn) + return 0; + return 1; +} + +/* Free optimized instruction slot */ +static void +__arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty) +{ + if (op->optinsn.insn) { + free_optinsn_slot(op->optinsn.insn, dirty); + op->optinsn.insn = NULL; + } +} + +extern void kprobe_handler(struct pt_regs *regs); + +static void +optimized_callback(struct optimized_kprobe *op, struct pt_regs *regs) +{ + unsigned long flags; + struct kprobe_ctlblk *kcb; + + if (kprobe_disabled(&op->kp)) + return; + + /* Save skipped registers */ + regs->pc = (unsigned long)op->kp.addr; + regs->orig_x0 = ~0UL; + regs->stackframe[1] = (unsigned long)op->kp.addr + 4; + + local_irq_save(flags); + kcb = get_kprobe_ctlblk(); + + if (kprobe_running()) { + kprobes_inc_nmissed_count(&op->kp); + } else { + __this_cpu_write(current_kprobe, &op->kp); + kcb->kprobe_status = KPROBE_HIT_ACTIVE; + opt_pre_handler(&op->kp, regs); + __this_cpu_write(current_kprobe, NULL); + } + + local_irq_restore(flags); +} +NOKPROBE_SYMBOL(optimized_callback) + +int arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *orig) +{ + kprobe_opcode_t *code; + void **addrs; + long offset; + kprobe_opcode_t final_branch; + u32 insns[8]; + int i; + + if (!can_optimize(orig)) + return -EILSEQ; + + /* Allocate instruction slot */ + code = get_optinsn_slot(); + if (!code) + return -ENOMEM; + + /* use a 'b' instruction to branch to optinsn.insn. + * according armv8 manual, branch range is +/-128MB, + * is encoded as "imm26" times 4. + * 31 30 26 + * +---+-----------+----------------+ + * | 0 | 0 0 1 0 1 | imm26 | + * +---+-----------+----------------+ + */ + offset = (long)code - (long)orig->addr; + + if (offset > 0x7ffffffL || offset < -0x8000000 || offset & 0x3) { + + free_optinsn_slot(code, 0); + return -ERANGE; + } + + addrs = kmalloc(MAX_OPTINSN_SIZE * sizeof(kprobe_opcode_t *), GFP_KERNEL); + for (i = 0; i < MAX_OPTINSN_SIZE; i++) + addrs[i] = &code[i]; + + /* Copy arch-dep-instance from template. */ + aarch64_insn_patch_text(addrs, + (kprobe_opcode_t *)&optprobe_template_entry, + TMPL_RESTORE_ORIGN_INSN); + + /* Set probe information */ + *(unsigned long *)&insns[TMPL_VAL_IDX-TMPL_RESTORE_ORIGN_INSN] = (unsigned long)op; + + + /* Set probe function call */ + *(unsigned long *)&insns[TMPL_CALL_IDX-TMPL_RESTORE_ORIGN_INSN] = (unsigned long)optimized_callback; + + final_branch = aarch64_insn_gen_branch_imm((unsigned long)(&code[TMPL_RESTORE_END]), + (unsigned long)(op->kp.addr) + 4, + AARCH64_INSN_BRANCH_NOLINK); + + /* The original probed instruction */ + if (orig->ainsn.api.insn) + insns[0] = orig->opcode; + else + insns[0] = 0xd503201f; /*nop*/ + + /* Jump back to next instruction */ + insns[1] = final_branch; + + aarch64_insn_patch_text(addrs + TMPL_RESTORE_ORIGN_INSN, + insns, + TMPL_END_IDX - TMPL_RESTORE_ORIGN_INSN); + + flush_icache_range((unsigned long)code, (unsigned long)(&code[TMPL_END_IDX])); + + /* Set op->optinsn.insn means prepared. */ + op->optinsn.insn = code; + + kfree(addrs); + + return 0; +} + +void __kprobes arch_optimize_kprobes(struct list_head *oplist) +{ + struct optimized_kprobe *op, *tmp; + + list_for_each_entry_safe(op, tmp, oplist, list) { + unsigned long insn; + void *addrs[] = {0}; + u32 insns[] = {0}; + + WARN_ON(kprobe_disabled(&op->kp)); + + /* + * Backup instructions which will be replaced + * by jump address + */ + memcpy(op->optinsn.copied_insn, op->kp.addr, + AARCH64_INSN_SIZE); + + insn = aarch64_insn_gen_branch_imm((unsigned long)op->kp.addr, + (unsigned long)op->optinsn.insn, + AARCH64_INSN_BRANCH_NOLINK); + + insns[0] = insn; + addrs[0] = op->kp.addr; + + aarch64_insn_patch_text(addrs, insns, 1); + + list_del_init(&op->list); + } +} + +void arch_unoptimize_kprobe(struct optimized_kprobe *op) +{ + arch_arm_kprobe(&op->kp); +} + +/* + * Recover original instructions and breakpoints from relative jumps. + * Caller must call with locking kprobe_mutex. + */ +void arch_unoptimize_kprobes(struct list_head *oplist, + struct list_head *done_list) +{ + struct optimized_kprobe *op, *tmp; + + list_for_each_entry_safe(op, tmp, oplist, list) { + arch_unoptimize_kprobe(op); + list_move(&op->list, done_list); + } +} + +int arch_within_optimized_kprobe(struct optimized_kprobe *op, + unsigned long addr) +{ + return ((unsigned long)op->kp.addr <= addr && + (unsigned long)op->kp.addr + AARCH64_INSN_SIZE > addr); +} + +void arch_remove_optimized_kprobe(struct optimized_kprobe *op) +{ + __arch_remove_optimized_kprobe(op, 1); +} diff --git a/arch/arm64/kernel/probes/opt_head.S b/arch/arm64/kernel/probes/opt_head.S new file mode 100644 index 0000000..0b437c3 --- /dev/null +++ b/arch/arm64/kernel/probes/opt_head.S @@ -0,0 +1,40 @@ +/* SPDX-License-Identifier: GPL-2.0-only + * Copyright 2021 Unisoc Inc. + */ +#include +#include +#include +#include "base_regs.h" + .align 2 + .global optinsn_slot + optinsn_slot: + .space 4096 + .global optprobe_template_entry + optprobe_template_entry: + + save_all_base_regs + + mov x1, sp + ldr x0, 1f + ldr x2, 2f + blr x2 + nop + + restore_all_base_regs 1 + + .global optprobe_template_restore_orig_insn + optprobe_template_restore_orig_insn: + nop + .global optprobe_template_restore_end + optprobe_template_restore_end: + nop + .align 3 + .global optprobe_template_val + optprobe_template_val: +1: .space 8 + .global optprobe_template_call + optprobe_template_call: +2: .space 8 + .global optprobe_template_end + optprobe_template_end: +