[PATCH] riscv/kprobe: Optimize the performance of patching instruction slot
Masami Hiramatsu (Google)
mhiramat at kernel.org
Fri Sep 9 19:24:35 PDT 2022
On Fri, 9 Sep 2022 09:55:08 +0800
"liaochang (A)" <liaochang1 at huawei.com> wrote:
>
>
> 在 2022/9/8 20:49, Masami Hiramatsu (Google) 写道:
> > On Thu, 8 Sep 2022 09:43:45 +0800
> > "liaochang (A)" <liaochang1 at huawei.com> wrote:
> >
> >> Thanks for comment.
> >>
> >> 在 2022/9/8 1:21, Jisheng Zhang 写道:
> >>> On Wed, Sep 07, 2022 at 10:33:27AM +0800, Liao Chang wrote:
> >>>> Since no race condition occurs on each instruction slot, hence it is
> >>>> safe to patch instruction slot without stopping machine.
> >>>
> >>> hmm, IMHO there's race when arming kprobe under SMP, so stopping
> >>> machine is necessary here. Maybe I misundertand something.
> >>>
> >>
> >> It is indeed necessary to stop machine when arm kprobe under SMP,
> >> but i don't think it need to stop machine when prepare instruction slot,
> >> two reasons:
> >>
> >> 1. Instruction slot is dynamically allocated data.
> >> 2. Kernel would not execute instruction slot until original instruction
> >> is replaced by breakpoint.
> >
> > Ah, this is for ss (single step out of line) slot. So until
> > kprobe is enabled, this should not be used from other cores.
> > OK, then it should be safe.
>
> Exactly, Masami, and i find out this optimization could be applied to some other
> architectures, such as arm64 and csky, do you think it is good time to do them all.
Yes, we should reduce the stop_machine() usage. Thanks for pointing it!
>
> Thanks.
>
> >
> >
> >>>>
> >>>> Signed-off-by: Liao Chang <liaochang1 at huawei.com>
> >>>> ---
> >>>> arch/riscv/kernel/probes/kprobes.c | 8 +++++---
> >>>> 1 file changed, 5 insertions(+), 3 deletions(-)
> >>>>
> >>>> diff --git a/arch/riscv/kernel/probes/kprobes.c b/arch/riscv/kernel/probes/kprobes.c
> >>>> index e6e950b7cf32..eff7d7fab535 100644
> >>>> --- a/arch/riscv/kernel/probes/kprobes.c
> >>>> +++ b/arch/riscv/kernel/probes/kprobes.c
> >>>> @@ -24,12 +24,14 @@ post_kprobe_handler(struct kprobe *, struct kprobe_ctlblk *, struct pt_regs *);
> >>>> static void __kprobes arch_prepare_ss_slot(struct kprobe *p)
> >>>> {
> >>>> unsigned long offset = GET_INSN_LENGTH(p->opcode);
> >>>> + const kprobe_opcode_t brk_insn = __BUG_INSN_32;
> >>>> + kprobe_opcode_t slot[MAX_INSN_SIZE];
> >>>>
> >>>> p->ainsn.api.restore = (unsigned long)p->addr + offset;
> >>>>
> >>>> - patch_text(p->ainsn.api.insn, p->opcode);
> >>>> - patch_text((void *)((unsigned long)(p->ainsn.api.insn) + offset),
> >>>> - __BUG_INSN_32);
> >>>> + memcpy(slot, &p->opcode, offset);
> >>>> + memcpy((void *)((unsigned long)slot + offset), &brk_insn, 4);
> >>>> + patch_text_nosync(p->ainsn.api.insn, slot, offset + 4);
> >
> > BTW, didn't you have a macro for the size of __BUG_INSN_32?
> >
> > Thank you,
>
> I think you are saying GET_INSN_LENGTH, i will use it to caculate
> the size of __BUG_INSN_32 in v2, instead of magic number '4'.
Yeah, that's better.
Thank you!
>
> Thanks.
>
> >
> >
> >>>> }
> >>>>
> >>>> static void __kprobes arch_prepare_simulate(struct kprobe *p)
> >>>> --
> >>>> 2.17.1
> >>>>
> >>>>
> >>>> _______________________________________________
> >>>> linux-riscv mailing list
> >>>> linux-riscv at lists.infradead.org
> >>>> http://lists.infradead.org/mailman/listinfo/linux-riscv
> >>> .
> >>
> >> --
> >> BR,
> >> Liao, Chang
> >
> >
>
> --
> BR,
> Liao, Chang
--
Masami Hiramatsu (Google) <mhiramat at kernel.org>
More information about the linux-riscv
mailing list