Linux Audio

Check our new training course

Loading...
v6.2
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * eBPF JIT compiler
  4 *
  5 * Copyright 2016 Naveen N. Rao <naveen.n.rao@linux.vnet.ibm.com>
  6 *		  IBM Corporation
  7 *
  8 * Based on the powerpc classic BPF JIT compiler by Matt Evans
  9 */
 10#include <linux/moduleloader.h>
 11#include <asm/cacheflush.h>
 12#include <asm/asm-compat.h>
 13#include <linux/netdevice.h>
 14#include <linux/filter.h>
 15#include <linux/if_vlan.h>
 16#include <asm/kprobes.h>
 17#include <linux/bpf.h>
 18
 19#include "bpf_jit.h"
 20
 21static void bpf_jit_fill_ill_insns(void *area, unsigned int size)
 22{
 23	memset32(area, BREAKPOINT_INSTRUCTION, size / 4);
 24}
 25
 26/* Fix updated addresses (for subprog calls, ldimm64, et al) during extra pass */
 27static int bpf_jit_fixup_addresses(struct bpf_prog *fp, u32 *image,
 28				   struct codegen_context *ctx, u32 *addrs)
 29{
 30	const struct bpf_insn *insn = fp->insnsi;
 31	bool func_addr_fixed;
 32	u64 func_addr;
 33	u32 tmp_idx;
 34	int i, j, ret;
 35
 36	for (i = 0; i < fp->len; i++) {
 37		/*
 38		 * During the extra pass, only the branch target addresses for
 39		 * the subprog calls need to be fixed. All other instructions
 40		 * can left untouched.
 41		 *
 42		 * The JITed image length does not change because we already
 43		 * ensure that the JITed instruction sequence for these calls
 44		 * are of fixed length by padding them with NOPs.
 45		 */
 46		if (insn[i].code == (BPF_JMP | BPF_CALL) &&
 47		    insn[i].src_reg == BPF_PSEUDO_CALL) {
 48			ret = bpf_jit_get_func_addr(fp, &insn[i], true,
 49						    &func_addr,
 50						    &func_addr_fixed);
 51			if (ret < 0)
 52				return ret;
 53
 54			/*
 55			 * Save ctx->idx as this would currently point to the
 56			 * end of the JITed image and set it to the offset of
 57			 * the instruction sequence corresponding to the
 58			 * subprog call temporarily.
 59			 */
 60			tmp_idx = ctx->idx;
 61			ctx->idx = addrs[i] / 4;
 62			ret = bpf_jit_emit_func_call_rel(image, ctx, func_addr);
 63			if (ret)
 64				return ret;
 65
 66			/*
 67			 * Restore ctx->idx here. This is safe as the length
 68			 * of the JITed sequence remains unchanged.
 69			 */
 70			ctx->idx = tmp_idx;
 71		} else if (insn[i].code == (BPF_LD | BPF_IMM | BPF_DW)) {
 72			tmp_idx = ctx->idx;
 73			ctx->idx = addrs[i] / 4;
 74#ifdef CONFIG_PPC32
 75			PPC_LI32(bpf_to_ppc(insn[i].dst_reg) - 1, (u32)insn[i + 1].imm);
 76			PPC_LI32(bpf_to_ppc(insn[i].dst_reg), (u32)insn[i].imm);
 77			for (j = ctx->idx - addrs[i] / 4; j < 4; j++)
 78				EMIT(PPC_RAW_NOP());
 79#else
 80			func_addr = ((u64)(u32)insn[i].imm) | (((u64)(u32)insn[i + 1].imm) << 32);
 81			PPC_LI64(bpf_to_ppc(insn[i].dst_reg), func_addr);
 82			/* overwrite rest with nops */
 83			for (j = ctx->idx - addrs[i] / 4; j < 5; j++)
 84				EMIT(PPC_RAW_NOP());
 85#endif
 86			ctx->idx = tmp_idx;
 87			i++;
 88		}
 89	}
 90
 91	return 0;
 92}
 93
 94int bpf_jit_emit_exit_insn(u32 *image, struct codegen_context *ctx, int tmp_reg, long exit_addr)
 95{
 96	if (!exit_addr || is_offset_in_branch_range(exit_addr - (ctx->idx * 4))) {
 97		PPC_JMP(exit_addr);
 98	} else if (ctx->alt_exit_addr) {
 99		if (WARN_ON(!is_offset_in_branch_range((long)ctx->alt_exit_addr - (ctx->idx * 4))))
100			return -1;
101		PPC_JMP(ctx->alt_exit_addr);
102	} else {
103		ctx->alt_exit_addr = ctx->idx * 4;
104		bpf_jit_build_epilogue(image, ctx);
105	}
106
107	return 0;
108}
109
110struct powerpc64_jit_data {
111	struct bpf_binary_header *header;
112	u32 *addrs;
113	u8 *image;
114	u32 proglen;
115	struct codegen_context ctx;
116};
117
118bool bpf_jit_needs_zext(void)
119{
120	return true;
121}
122
123struct bpf_prog *bpf_int_jit_compile(struct bpf_prog *fp)
124{
125	u32 proglen;
126	u32 alloclen;
127	u8 *image = NULL;
128	u32 *code_base;
129	u32 *addrs;
130	struct powerpc64_jit_data *jit_data;
131	struct codegen_context cgctx;
132	int pass;
133	int flen;
134	struct bpf_binary_header *bpf_hdr;
135	struct bpf_prog *org_fp = fp;
136	struct bpf_prog *tmp_fp;
137	bool bpf_blinded = false;
138	bool extra_pass = false;
139	u32 extable_len;
140	u32 fixup_len;
141
142	if (!fp->jit_requested)
143		return org_fp;
144
145	tmp_fp = bpf_jit_blind_constants(org_fp);
146	if (IS_ERR(tmp_fp))
147		return org_fp;
148
149	if (tmp_fp != org_fp) {
150		bpf_blinded = true;
151		fp = tmp_fp;
152	}
153
154	jit_data = fp->aux->jit_data;
155	if (!jit_data) {
156		jit_data = kzalloc(sizeof(*jit_data), GFP_KERNEL);
157		if (!jit_data) {
158			fp = org_fp;
159			goto out;
160		}
161		fp->aux->jit_data = jit_data;
162	}
163
164	flen = fp->len;
165	addrs = jit_data->addrs;
166	if (addrs) {
167		cgctx = jit_data->ctx;
168		image = jit_data->image;
169		bpf_hdr = jit_data->header;
170		proglen = jit_data->proglen;
 
171		extra_pass = true;
172		goto skip_init_ctx;
173	}
174
175	addrs = kcalloc(flen + 1, sizeof(*addrs), GFP_KERNEL);
176	if (addrs == NULL) {
177		fp = org_fp;
178		goto out_addrs;
179	}
180
181	memset(&cgctx, 0, sizeof(struct codegen_context));
182	bpf_jit_init_reg_mapping(&cgctx);
183
184	/* Make sure that the stack is quadword aligned. */
185	cgctx.stack_size = round_up(fp->aux->stack_depth, 16);
186
187	/* Scouting faux-generate pass 0 */
188	if (bpf_jit_build_body(fp, 0, &cgctx, addrs, 0)) {
189		/* We hit something illegal or unsupported. */
190		fp = org_fp;
191		goto out_addrs;
192	}
193
194	/*
195	 * If we have seen a tail call, we need a second pass.
196	 * This is because bpf_jit_emit_common_epilogue() is called
197	 * from bpf_jit_emit_tail_call() with a not yet stable ctx->seen.
198	 * We also need a second pass if we ended up with too large
199	 * a program so as to ensure BPF_EXIT branches are in range.
200	 */
201	if (cgctx.seen & SEEN_TAILCALL || !is_offset_in_branch_range((long)cgctx.idx * 4)) {
202		cgctx.idx = 0;
203		if (bpf_jit_build_body(fp, 0, &cgctx, addrs, 0)) {
204			fp = org_fp;
205			goto out_addrs;
206		}
207	}
208
209	bpf_jit_realloc_regs(&cgctx);
210	/*
211	 * Pretend to build prologue, given the features we've seen.  This will
212	 * update ctgtx.idx as it pretends to output instructions, then we can
213	 * calculate total size from idx.
214	 */
215	bpf_jit_build_prologue(0, &cgctx);
216	addrs[fp->len] = cgctx.idx * 4;
217	bpf_jit_build_epilogue(0, &cgctx);
218
219	fixup_len = fp->aux->num_exentries * BPF_FIXUP_LEN * 4;
220	extable_len = fp->aux->num_exentries * sizeof(struct exception_table_entry);
221
222	proglen = cgctx.idx * 4;
223	alloclen = proglen + FUNCTION_DESCR_SIZE + fixup_len + extable_len;
224
225	bpf_hdr = bpf_jit_binary_alloc(alloclen, &image, 4, bpf_jit_fill_ill_insns);
226	if (!bpf_hdr) {
227		fp = org_fp;
228		goto out_addrs;
229	}
230
231	if (extable_len)
232		fp->aux->extable = (void *)image + FUNCTION_DESCR_SIZE + proglen + fixup_len;
233
234skip_init_ctx:
235	code_base = (u32 *)(image + FUNCTION_DESCR_SIZE);
236
237	if (extra_pass) {
238		/*
239		 * Do not touch the prologue and epilogue as they will remain
240		 * unchanged. Only fix the branch target address for subprog
241		 * calls in the body, and ldimm64 instructions.
242		 *
243		 * This does not change the offsets and lengths of the subprog
244		 * call instruction sequences and hence, the size of the JITed
245		 * image as well.
246		 */
247		bpf_jit_fixup_addresses(fp, code_base, &cgctx, addrs);
248
249		/* There is no need to perform the usual passes. */
250		goto skip_codegen_passes;
251	}
252
253	/* Code generation passes 1-2 */
254	for (pass = 1; pass < 3; pass++) {
255		/* Now build the prologue, body code & epilogue for real. */
256		cgctx.idx = 0;
257		cgctx.alt_exit_addr = 0;
258		bpf_jit_build_prologue(code_base, &cgctx);
259		if (bpf_jit_build_body(fp, code_base, &cgctx, addrs, pass)) {
260			bpf_jit_binary_free(bpf_hdr);
261			fp = org_fp;
262			goto out_addrs;
263		}
264		bpf_jit_build_epilogue(code_base, &cgctx);
265
266		if (bpf_jit_enable > 1)
267			pr_info("Pass %d: shrink = %d, seen = 0x%x\n", pass,
268				proglen - (cgctx.idx * 4), cgctx.seen);
269	}
270
271skip_codegen_passes:
272	if (bpf_jit_enable > 1)
273		/*
274		 * Note that we output the base address of the code_base
275		 * rather than image, since opcodes are in code_base.
276		 */
277		bpf_jit_dump(flen, proglen, pass, code_base);
278
279#ifdef CONFIG_PPC64_ELF_ABI_V1
280	/* Function descriptor nastiness: Address + TOC */
281	((u64 *)image)[0] = (u64)code_base;
282	((u64 *)image)[1] = local_paca->kernel_toc;
283#endif
284
285	fp->bpf_func = (void *)image;
286	fp->jited = 1;
287	fp->jited_len = proglen + FUNCTION_DESCR_SIZE;
288
289	bpf_flush_icache(bpf_hdr, (u8 *)bpf_hdr + bpf_hdr->size);
 
290	if (!fp->is_func || extra_pass) {
291		bpf_jit_binary_lock_ro(bpf_hdr);
292		bpf_prog_fill_jited_linfo(fp, addrs);
293out_addrs:
294		kfree(addrs);
295		kfree(jit_data);
296		fp->aux->jit_data = NULL;
297	} else {
298		jit_data->addrs = addrs;
299		jit_data->ctx = cgctx;
300		jit_data->proglen = proglen;
301		jit_data->image = image;
302		jit_data->header = bpf_hdr;
303	}
304
305out:
306	if (bpf_blinded)
307		bpf_jit_prog_release_other(fp, fp == org_fp ? tmp_fp : org_fp);
308
309	return fp;
310}
311
312/*
313 * The caller should check for (BPF_MODE(code) == BPF_PROBE_MEM) before calling
314 * this function, as this only applies to BPF_PROBE_MEM, for now.
315 */
316int bpf_add_extable_entry(struct bpf_prog *fp, u32 *image, int pass, struct codegen_context *ctx,
317			  int insn_idx, int jmp_off, int dst_reg)
318{
319	off_t offset;
320	unsigned long pc;
321	struct exception_table_entry *ex;
322	u32 *fixup;
323
324	/* Populate extable entries only in the last pass */
325	if (pass != 2)
326		return 0;
327
328	if (!fp->aux->extable ||
329	    WARN_ON_ONCE(ctx->exentry_idx >= fp->aux->num_exentries))
330		return -EINVAL;
331
332	pc = (unsigned long)&image[insn_idx];
333
334	fixup = (void *)fp->aux->extable -
335		(fp->aux->num_exentries * BPF_FIXUP_LEN * 4) +
336		(ctx->exentry_idx * BPF_FIXUP_LEN * 4);
337
338	fixup[0] = PPC_RAW_LI(dst_reg, 0);
339	if (IS_ENABLED(CONFIG_PPC32))
340		fixup[1] = PPC_RAW_LI(dst_reg - 1, 0); /* clear higher 32-bit register too */
341
342	fixup[BPF_FIXUP_LEN - 1] =
343		PPC_RAW_BRANCH((long)(pc + jmp_off) - (long)&fixup[BPF_FIXUP_LEN - 1]);
344
345	ex = &fp->aux->extable[ctx->exentry_idx];
346
347	offset = pc - (long)&ex->insn;
348	if (WARN_ON_ONCE(offset >= 0 || offset < INT_MIN))
349		return -ERANGE;
350	ex->insn = offset;
351
352	offset = (long)fixup - (long)&ex->fixup;
353	if (WARN_ON_ONCE(offset >= 0 || offset < INT_MIN))
354		return -ERANGE;
355	ex->fixup = offset;
356
357	ctx->exentry_idx++;
358	return 0;
359}
v5.14.15
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * eBPF JIT compiler
  4 *
  5 * Copyright 2016 Naveen N. Rao <naveen.n.rao@linux.vnet.ibm.com>
  6 *		  IBM Corporation
  7 *
  8 * Based on the powerpc classic BPF JIT compiler by Matt Evans
  9 */
 10#include <linux/moduleloader.h>
 11#include <asm/cacheflush.h>
 12#include <asm/asm-compat.h>
 13#include <linux/netdevice.h>
 14#include <linux/filter.h>
 15#include <linux/if_vlan.h>
 16#include <asm/kprobes.h>
 17#include <linux/bpf.h>
 18
 19#include "bpf_jit.h"
 20
 21static void bpf_jit_fill_ill_insns(void *area, unsigned int size)
 22{
 23	memset32(area, BREAKPOINT_INSTRUCTION, size / 4);
 24}
 25
 26/* Fix the branch target addresses for subprog calls */
 27static int bpf_jit_fixup_subprog_calls(struct bpf_prog *fp, u32 *image,
 28				       struct codegen_context *ctx, u32 *addrs)
 29{
 30	const struct bpf_insn *insn = fp->insnsi;
 31	bool func_addr_fixed;
 32	u64 func_addr;
 33	u32 tmp_idx;
 34	int i, ret;
 35
 36	for (i = 0; i < fp->len; i++) {
 37		/*
 38		 * During the extra pass, only the branch target addresses for
 39		 * the subprog calls need to be fixed. All other instructions
 40		 * can left untouched.
 41		 *
 42		 * The JITed image length does not change because we already
 43		 * ensure that the JITed instruction sequence for these calls
 44		 * are of fixed length by padding them with NOPs.
 45		 */
 46		if (insn[i].code == (BPF_JMP | BPF_CALL) &&
 47		    insn[i].src_reg == BPF_PSEUDO_CALL) {
 48			ret = bpf_jit_get_func_addr(fp, &insn[i], true,
 49						    &func_addr,
 50						    &func_addr_fixed);
 51			if (ret < 0)
 52				return ret;
 53
 54			/*
 55			 * Save ctx->idx as this would currently point to the
 56			 * end of the JITed image and set it to the offset of
 57			 * the instruction sequence corresponding to the
 58			 * subprog call temporarily.
 59			 */
 60			tmp_idx = ctx->idx;
 61			ctx->idx = addrs[i] / 4;
 62			bpf_jit_emit_func_call_rel(image, ctx, func_addr);
 
 
 63
 64			/*
 65			 * Restore ctx->idx here. This is safe as the length
 66			 * of the JITed sequence remains unchanged.
 67			 */
 68			ctx->idx = tmp_idx;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 69		}
 70	}
 71
 72	return 0;
 73}
 74
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 75struct powerpc64_jit_data {
 76	struct bpf_binary_header *header;
 77	u32 *addrs;
 78	u8 *image;
 79	u32 proglen;
 80	struct codegen_context ctx;
 81};
 82
 83bool bpf_jit_needs_zext(void)
 84{
 85	return true;
 86}
 87
 88struct bpf_prog *bpf_int_jit_compile(struct bpf_prog *fp)
 89{
 90	u32 proglen;
 91	u32 alloclen;
 92	u8 *image = NULL;
 93	u32 *code_base;
 94	u32 *addrs;
 95	struct powerpc64_jit_data *jit_data;
 96	struct codegen_context cgctx;
 97	int pass;
 98	int flen;
 99	struct bpf_binary_header *bpf_hdr;
100	struct bpf_prog *org_fp = fp;
101	struct bpf_prog *tmp_fp;
102	bool bpf_blinded = false;
103	bool extra_pass = false;
 
 
104
105	if (!fp->jit_requested)
106		return org_fp;
107
108	tmp_fp = bpf_jit_blind_constants(org_fp);
109	if (IS_ERR(tmp_fp))
110		return org_fp;
111
112	if (tmp_fp != org_fp) {
113		bpf_blinded = true;
114		fp = tmp_fp;
115	}
116
117	jit_data = fp->aux->jit_data;
118	if (!jit_data) {
119		jit_data = kzalloc(sizeof(*jit_data), GFP_KERNEL);
120		if (!jit_data) {
121			fp = org_fp;
122			goto out;
123		}
124		fp->aux->jit_data = jit_data;
125	}
126
127	flen = fp->len;
128	addrs = jit_data->addrs;
129	if (addrs) {
130		cgctx = jit_data->ctx;
131		image = jit_data->image;
132		bpf_hdr = jit_data->header;
133		proglen = jit_data->proglen;
134		alloclen = proglen + FUNCTION_DESCR_SIZE;
135		extra_pass = true;
136		goto skip_init_ctx;
137	}
138
139	addrs = kcalloc(flen + 1, sizeof(*addrs), GFP_KERNEL);
140	if (addrs == NULL) {
141		fp = org_fp;
142		goto out_addrs;
143	}
144
145	memset(&cgctx, 0, sizeof(struct codegen_context));
146	memcpy(cgctx.b2p, b2p, sizeof(cgctx.b2p));
147
148	/* Make sure that the stack is quadword aligned. */
149	cgctx.stack_size = round_up(fp->aux->stack_depth, 16);
150
151	/* Scouting faux-generate pass 0 */
152	if (bpf_jit_build_body(fp, 0, &cgctx, addrs, false)) {
153		/* We hit something illegal or unsupported. */
154		fp = org_fp;
155		goto out_addrs;
156	}
157
158	/*
159	 * If we have seen a tail call, we need a second pass.
160	 * This is because bpf_jit_emit_common_epilogue() is called
161	 * from bpf_jit_emit_tail_call() with a not yet stable ctx->seen.
 
 
162	 */
163	if (cgctx.seen & SEEN_TAILCALL) {
164		cgctx.idx = 0;
165		if (bpf_jit_build_body(fp, 0, &cgctx, addrs, false)) {
166			fp = org_fp;
167			goto out_addrs;
168		}
169	}
170
171	bpf_jit_realloc_regs(&cgctx);
172	/*
173	 * Pretend to build prologue, given the features we've seen.  This will
174	 * update ctgtx.idx as it pretends to output instructions, then we can
175	 * calculate total size from idx.
176	 */
177	bpf_jit_build_prologue(0, &cgctx);
 
178	bpf_jit_build_epilogue(0, &cgctx);
179
 
 
 
180	proglen = cgctx.idx * 4;
181	alloclen = proglen + FUNCTION_DESCR_SIZE;
182
183	bpf_hdr = bpf_jit_binary_alloc(alloclen, &image, 4, bpf_jit_fill_ill_insns);
184	if (!bpf_hdr) {
185		fp = org_fp;
186		goto out_addrs;
187	}
188
 
 
 
189skip_init_ctx:
190	code_base = (u32 *)(image + FUNCTION_DESCR_SIZE);
191
192	if (extra_pass) {
193		/*
194		 * Do not touch the prologue and epilogue as they will remain
195		 * unchanged. Only fix the branch target address for subprog
196		 * calls in the body.
197		 *
198		 * This does not change the offsets and lengths of the subprog
199		 * call instruction sequences and hence, the size of the JITed
200		 * image as well.
201		 */
202		bpf_jit_fixup_subprog_calls(fp, code_base, &cgctx, addrs);
203
204		/* There is no need to perform the usual passes. */
205		goto skip_codegen_passes;
206	}
207
208	/* Code generation passes 1-2 */
209	for (pass = 1; pass < 3; pass++) {
210		/* Now build the prologue, body code & epilogue for real. */
211		cgctx.idx = 0;
 
212		bpf_jit_build_prologue(code_base, &cgctx);
213		if (bpf_jit_build_body(fp, code_base, &cgctx, addrs, extra_pass)) {
214			bpf_jit_binary_free(bpf_hdr);
215			fp = org_fp;
216			goto out_addrs;
217		}
218		bpf_jit_build_epilogue(code_base, &cgctx);
219
220		if (bpf_jit_enable > 1)
221			pr_info("Pass %d: shrink = %d, seen = 0x%x\n", pass,
222				proglen - (cgctx.idx * 4), cgctx.seen);
223	}
224
225skip_codegen_passes:
226	if (bpf_jit_enable > 1)
227		/*
228		 * Note that we output the base address of the code_base
229		 * rather than image, since opcodes are in code_base.
230		 */
231		bpf_jit_dump(flen, proglen, pass, code_base);
232
233#ifdef PPC64_ELF_ABI_v1
234	/* Function descriptor nastiness: Address + TOC */
235	((u64 *)image)[0] = (u64)code_base;
236	((u64 *)image)[1] = local_paca->kernel_toc;
237#endif
238
239	fp->bpf_func = (void *)image;
240	fp->jited = 1;
241	fp->jited_len = alloclen;
242
243	bpf_flush_icache(bpf_hdr, (u8 *)bpf_hdr + (bpf_hdr->pages * PAGE_SIZE));
244	bpf_jit_binary_lock_ro(bpf_hdr);
245	if (!fp->is_func || extra_pass) {
 
246		bpf_prog_fill_jited_linfo(fp, addrs);
247out_addrs:
248		kfree(addrs);
249		kfree(jit_data);
250		fp->aux->jit_data = NULL;
251	} else {
252		jit_data->addrs = addrs;
253		jit_data->ctx = cgctx;
254		jit_data->proglen = proglen;
255		jit_data->image = image;
256		jit_data->header = bpf_hdr;
257	}
258
259out:
260	if (bpf_blinded)
261		bpf_jit_prog_release_other(fp, fp == org_fp ? tmp_fp : org_fp);
262
263	return fp;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
264}