On 02/03/2019 12:51 PM, bjorn.to...@gmail.com wrote: > From: Björn Töpel <bjorn.to...@gmail.com> > > This commit adds BPF JIT for RV64G. > > The JIT is a two-pass JIT, and has a dynamic prolog/epilogue (similar > to the MIPS64 BPF JIT) instead of static ones (e.g. x86_64). > > At the moment the RISC-V Linux port does not support HAVE_KPROBES, > which means that CONFIG_BPF_EVENTS is not supported. Thus, no tests > involving BPF_PROG_TYPE_TRACEPOINT passes. > > Further, the implementation does not support "far branching" (>4KiB). > > The implementation passes all the test_bpf.ko tests: > test_bpf: Summary: 378 PASSED, 0 FAILED, [366/366 JIT'ed] > > All the tail_call tests in the selftest/bpf/test_verifier program > passes. > > All tests where done on QEMU (QEMU emulator version 3.1.50 > (v3.1.0-688-g8ae951fbc106)). > > Signed-off-by: Björn Töpel <bjorn.to...@gmail.com>
Some minor comments: Looks like all the BPF_JMP32 instructions are missing. Would probably make sense to include these into the initial merge as well unless there is some good reason not to; presumably the test_verifier parts with BPF_JMP32 haven't been tried out? [...] > + > +enum { > + RV_CTX_F_SEEN_TAIL_CALL = 0, > + RV_CTX_F_SEEN_CALL = RV_REG_RA, > + RV_CTX_F_SEEN_S1 = RV_REG_S1, > + RV_CTX_F_SEEN_S2 = RV_REG_S2, > + RV_CTX_F_SEEN_S3 = RV_REG_S3, > + RV_CTX_F_SEEN_S4 = RV_REG_S4, > + RV_CTX_F_SEEN_S5 = RV_REG_S5, > + RV_CTX_F_SEEN_S6 = RV_REG_S6, > +}; > + > +struct rv_jit_context { > + struct bpf_prog *prog; > + u32 *insns; /* RV insns */ > + int ninsns; > + int epilogue_offset; > + int *offset; /* BPF to RV */ > + unsigned long flags; > + int stack_size; > +}; > + > +struct rv_jit_data { > + struct bpf_binary_header *header; > + u8 *image; > + struct rv_jit_context ctx; > +}; > + > +static u8 bpf_to_rv_reg(int bpf_reg, struct rv_jit_context *ctx) > +{ > + u8 reg = regmap[bpf_reg]; > + > + switch (reg) { > + case RV_CTX_F_SEEN_S1: > + case RV_CTX_F_SEEN_S2: > + case RV_CTX_F_SEEN_S3: > + case RV_CTX_F_SEEN_S4: > + case RV_CTX_F_SEEN_S5: > + case RV_CTX_F_SEEN_S6: > + __set_bit(reg, &ctx->flags); > + } > + return reg; > +}; > + > +static bool seen_reg(int reg, struct rv_jit_context *ctx) > +{ > + switch (reg) { > + case RV_CTX_F_SEEN_CALL: > + case RV_CTX_F_SEEN_S1: > + case RV_CTX_F_SEEN_S2: > + case RV_CTX_F_SEEN_S3: > + case RV_CTX_F_SEEN_S4: > + case RV_CTX_F_SEEN_S5: > + case RV_CTX_F_SEEN_S6: > + return test_bit(reg, &ctx->flags); > + } > + return false; > +} > + > +static void mark_call(struct rv_jit_context *ctx) > +{ > + __set_bit(RV_CTX_F_SEEN_CALL, &ctx->flags); > +} > + > +static bool seen_call(struct rv_jit_context *ctx) > +{ > + return seen_reg(RV_REG_RA, ctx); > +} Just nit: probably might be more obvious to remove this asymmetry in seen_reg() and do __set_bit()/test_bit() for RV_CTX_F_SEEN_CALL similar like below. > +static void mark_tail_call(struct rv_jit_context *ctx) > +{ > + __set_bit(RV_CTX_F_SEEN_TAIL_CALL, &ctx->flags); > +} > + > +static bool seen_tail_call(struct rv_jit_context *ctx) > +{ > + return test_bit(RV_CTX_F_SEEN_TAIL_CALL, &ctx->flags); > +} > + > +static u8 rv_tail_call_reg(struct rv_jit_context *ctx) > +{ > + mark_tail_call(ctx); > + > + if (seen_call(ctx)) { > + __set_bit(RV_CTX_F_SEEN_S6, &ctx->flags); > + return RV_REG_S6; > + } > + return RV_REG_A6; > +} > + > +static void emit(const u32 insn, struct rv_jit_context *ctx) > +{ > + if (ctx->insns) > + ctx->insns[ctx->ninsns] = insn; > + > + ctx->ninsns++; > +} > + > +static u32 rv_r_insn(u8 funct7, u8 rs2, u8 rs1, u8 funct3, u8 rd, u8 opcode) > +{ [...] > + /* Allocate image, now that we know the size. */ > + image_size = sizeof(u32) * ctx->ninsns; > + jit_data->header = bpf_jit_binary_alloc(image_size, &jit_data->image, > + sizeof(u32), > + bpf_fill_ill_insns); > + if (!jit_data->header) { > + prog = orig_prog; > + goto out_offset; > + } > + > + /* Second, real pass, that acutally emits the image. */ > + ctx->insns = (u32 *)jit_data->image; > +skip_init_ctx: > + ctx->ninsns = 0; > + > + build_prologue(ctx); > + if (build_body(ctx, extra_pass)) { > + bpf_jit_binary_free(jit_data->header); > + prog = orig_prog; > + goto out_offset; > + } > + build_epilogue(ctx); > + > + if (bpf_jit_enable > 1) > + bpf_jit_dump(prog->len, image_size, 2, ctx->insns); > + > + prog->bpf_func = (void *)ctx->insns; > + prog->jited = 1; > + prog->jited_len = image_size; > + > + bpf_flush_icache(jit_data->header, (u8 *)ctx->insns + ctx->ninsns); Shouldn't this be '(u32 *)ctx->insns + ctx->ninsns' to cover the range? > + > + if (!prog->is_func || extra_pass) { > +out_offset: > + kfree(ctx->offset); > + kfree(jit_data); > + prog->aux->jit_data = NULL; > + } > +out: > + if (tmp_blinded) > + bpf_jit_prog_release_other(prog, prog == orig_prog ? > + tmp : orig_prog); > + return prog; > +} >