1 // SPDX-License-Identifier: GPL-2.0-only 2 /* Copyright (c) 2026 Meta Platforms, Inc. and affiliates. */ 3 #include <linux/bpf.h> 4 #include <linux/btf.h> 5 #include <linux/bpf_verifier.h> 6 #include <linux/filter.h> 7 #include <linux/vmalloc.h> 8 #include <linux/bsearch.h> 9 #include <linux/sort.h> 10 #include <linux/perf_event.h> 11 #include <net/xdp.h> 12 #include "disasm.h" 13 14 #define verbose(env, fmt, args...) bpf_verifier_log_write(env, fmt, ##args) 15 16 static bool is_cmpxchg_insn(const struct bpf_insn *insn) 17 { 18 return BPF_CLASS(insn->code) == BPF_STX && 19 BPF_MODE(insn->code) == BPF_ATOMIC && 20 insn->imm == BPF_CMPXCHG; 21 } 22 23 /* Return the regno defined by the insn, or -1. */ 24 static int insn_def_regno(const struct bpf_insn *insn) 25 { 26 switch (BPF_CLASS(insn->code)) { 27 case BPF_JMP: 28 case BPF_JMP32: 29 case BPF_ST: 30 return -1; 31 case BPF_STX: 32 if (BPF_MODE(insn->code) == BPF_ATOMIC || 33 BPF_MODE(insn->code) == BPF_PROBE_ATOMIC) { 34 if (insn->imm == BPF_CMPXCHG) 35 return BPF_REG_0; 36 else if (insn->imm == BPF_LOAD_ACQ) 37 return insn->dst_reg; 38 else if (insn->imm & BPF_FETCH) 39 return insn->src_reg; 40 } 41 return -1; 42 default: 43 return insn->dst_reg; 44 } 45 } 46 47 /* Return TRUE if INSN has defined any 32-bit value explicitly. */ 48 static bool insn_has_def32(struct bpf_insn *insn) 49 { 50 int dst_reg = insn_def_regno(insn); 51 52 if (dst_reg == -1) 53 return false; 54 55 return !bpf_is_reg64(insn, dst_reg, NULL, DST_OP); 56 } 57 58 static int kfunc_desc_cmp_by_imm_off(const void *a, const void *b) 59 { 60 const struct bpf_kfunc_desc *d0 = a; 61 const struct bpf_kfunc_desc *d1 = b; 62 63 if (d0->imm != d1->imm) 64 return d0->imm < d1->imm ? -1 : 1; 65 if (d0->offset != d1->offset) 66 return d0->offset < d1->offset ? -1 : 1; 67 return 0; 68 } 69 70 const struct btf_func_model * 71 bpf_jit_find_kfunc_model(const struct bpf_prog *prog, 72 const struct bpf_insn *insn) 73 { 74 const struct bpf_kfunc_desc desc = { 75 .imm = insn->imm, 76 .offset = insn->off, 77 }; 78 const struct bpf_kfunc_desc *res; 79 struct bpf_kfunc_desc_tab *tab; 80 81 tab = prog->aux->kfunc_tab; 82 res = bsearch(&desc, tab->descs, tab->nr_descs, 83 sizeof(tab->descs[0]), kfunc_desc_cmp_by_imm_off); 84 85 return res ? &res->func_model : NULL; 86 } 87 88 static int set_kfunc_desc_imm(struct bpf_verifier_env *env, struct bpf_kfunc_desc *desc) 89 { 90 unsigned long call_imm; 91 92 if (bpf_jit_supports_far_kfunc_call()) { 93 call_imm = desc->func_id; 94 } else { 95 call_imm = BPF_CALL_IMM(desc->addr); 96 /* Check whether the relative offset overflows desc->imm */ 97 if ((unsigned long)(s32)call_imm != call_imm) { 98 verbose(env, "address of kernel func_id %u is out of range\n", 99 desc->func_id); 100 return -EINVAL; 101 } 102 } 103 desc->imm = call_imm; 104 return 0; 105 } 106 107 static int sort_kfunc_descs_by_imm_off(struct bpf_verifier_env *env) 108 { 109 struct bpf_kfunc_desc_tab *tab; 110 int i, err; 111 112 tab = env->prog->aux->kfunc_tab; 113 if (!tab) 114 return 0; 115 116 for (i = 0; i < tab->nr_descs; i++) { 117 err = set_kfunc_desc_imm(env, &tab->descs[i]); 118 if (err) 119 return err; 120 } 121 122 sort(tab->descs, tab->nr_descs, sizeof(tab->descs[0]), 123 kfunc_desc_cmp_by_imm_off, NULL); 124 return 0; 125 } 126 127 static int add_kfunc_in_insns(struct bpf_verifier_env *env, 128 struct bpf_insn *insn, int cnt) 129 { 130 int i, ret; 131 132 for (i = 0; i < cnt; i++, insn++) { 133 if (bpf_pseudo_kfunc_call(insn)) { 134 ret = bpf_add_kfunc_call(env, insn->imm, insn->off); 135 if (ret < 0) 136 return ret; 137 } 138 } 139 return 0; 140 } 141 142 #ifndef CONFIG_BPF_JIT_ALWAYS_ON 143 static int get_callee_stack_depth(struct bpf_verifier_env *env, 144 const struct bpf_insn *insn, int idx) 145 { 146 int start = idx + insn->imm + 1, subprog; 147 148 subprog = bpf_find_subprog(env, start); 149 if (verifier_bug_if(subprog < 0, env, "get stack depth: no program at insn %d", start)) 150 return -EFAULT; 151 return env->subprog_info[subprog].stack_depth; 152 } 153 #endif 154 155 /* single env->prog->insni[off] instruction was replaced with the range 156 * insni[off, off + cnt). Adjust corresponding insn_aux_data by copying 157 * [0, off) and [off, end) to new locations, so the patched range stays zero 158 */ 159 static void adjust_insn_aux_data(struct bpf_verifier_env *env, 160 struct bpf_prog *new_prog, u32 off, u32 cnt) 161 { 162 struct bpf_insn_aux_data *data = env->insn_aux_data; 163 struct bpf_insn *insn = new_prog->insnsi; 164 u32 old_seen = data[off].seen; 165 u32 prog_len; 166 int i; 167 168 /* aux info at OFF always needs adjustment, no matter fast path 169 * (cnt == 1) is taken or not. There is no guarantee INSN at OFF is the 170 * original insn at old prog. 171 */ 172 data[off].zext_dst = insn_has_def32(insn + off + cnt - 1); 173 174 if (cnt == 1) 175 return; 176 prog_len = new_prog->len; 177 178 memmove(data + off + cnt - 1, data + off, 179 sizeof(struct bpf_insn_aux_data) * (prog_len - off - cnt + 1)); 180 memset(data + off, 0, sizeof(struct bpf_insn_aux_data) * (cnt - 1)); 181 for (i = off; i < off + cnt - 1; i++) { 182 /* Expand insni[off]'s seen count to the patched range. */ 183 data[i].seen = old_seen; 184 data[i].zext_dst = insn_has_def32(insn + i); 185 } 186 187 /* 188 * The indirect_target flag of the original instruction was moved to the last of the 189 * new instructions by the above memmove and memset, but the indirect jump target is 190 * actually the first instruction, so move it back. This also matches with the behavior 191 * of bpf_insn_array_adjust(), which preserves xlated_off to point to the first new 192 * instruction. 193 */ 194 if (data[off + cnt - 1].indirect_target) { 195 data[off].indirect_target = 1; 196 data[off + cnt - 1].indirect_target = 0; 197 } 198 } 199 200 static void adjust_subprog_starts(struct bpf_verifier_env *env, u32 off, u32 len) 201 { 202 int i; 203 204 if (len == 1) 205 return; 206 /* NOTE: fake 'exit' subprog should be updated as well. */ 207 for (i = 0; i <= env->subprog_cnt; i++) { 208 if (env->subprog_info[i].start <= off) 209 continue; 210 env->subprog_info[i].start += len - 1; 211 } 212 } 213 214 static void adjust_insn_arrays(struct bpf_verifier_env *env, u32 off, u32 len) 215 { 216 int i; 217 218 if (len == 1) 219 return; 220 221 for (i = 0; i < env->insn_array_map_cnt; i++) 222 bpf_insn_array_adjust(env->insn_array_maps[i], off, len); 223 } 224 225 static void adjust_insn_arrays_after_remove(struct bpf_verifier_env *env, u32 off, u32 len) 226 { 227 int i; 228 229 for (i = 0; i < env->insn_array_map_cnt; i++) 230 bpf_insn_array_adjust_after_remove(env->insn_array_maps[i], off, len); 231 } 232 233 static void adjust_poke_descs(struct bpf_prog *prog, u32 off, u32 len) 234 { 235 struct bpf_jit_poke_descriptor *tab = prog->aux->poke_tab; 236 int i, sz = prog->aux->size_poke_tab; 237 struct bpf_jit_poke_descriptor *desc; 238 239 for (i = 0; i < sz; i++) { 240 desc = &tab[i]; 241 if (desc->insn_idx <= off) 242 continue; 243 desc->insn_idx += len - 1; 244 } 245 } 246 247 struct bpf_prog *bpf_patch_insn_data(struct bpf_verifier_env *env, u32 off, 248 const struct bpf_insn *patch, u32 len) 249 { 250 struct bpf_prog *new_prog; 251 struct bpf_insn_aux_data *new_data = NULL; 252 253 if (len > 1) { 254 new_data = vrealloc(env->insn_aux_data, 255 array_size(env->prog->len + len - 1, 256 sizeof(struct bpf_insn_aux_data)), 257 GFP_KERNEL_ACCOUNT | __GFP_ZERO); 258 if (!new_data) 259 return NULL; 260 261 env->insn_aux_data = new_data; 262 } 263 264 new_prog = bpf_patch_insn_single(env->prog, off, patch, len); 265 if (IS_ERR(new_prog)) { 266 if (PTR_ERR(new_prog) == -ERANGE) 267 verbose(env, 268 "insn %d cannot be patched due to 16-bit range\n", 269 env->insn_aux_data[off].orig_idx); 270 return NULL; 271 } 272 adjust_insn_aux_data(env, new_prog, off, len); 273 adjust_subprog_starts(env, off, len); 274 adjust_insn_arrays(env, off, len); 275 adjust_poke_descs(new_prog, off, len); 276 return new_prog; 277 } 278 279 /* 280 * For all jmp insns in a given 'prog' that point to 'tgt_idx' insn adjust the 281 * jump offset by 'delta'. 282 */ 283 static int adjust_jmp_off(struct bpf_prog *prog, u32 tgt_idx, u32 delta) 284 { 285 struct bpf_insn *insn = prog->insnsi; 286 u32 insn_cnt = prog->len, i; 287 s32 imm; 288 s16 off; 289 290 for (i = 0; i < insn_cnt; i++, insn++) { 291 u8 code = insn->code; 292 293 if (tgt_idx <= i && i < tgt_idx + delta) 294 continue; 295 296 if ((BPF_CLASS(code) != BPF_JMP && BPF_CLASS(code) != BPF_JMP32) || 297 BPF_OP(code) == BPF_CALL || BPF_OP(code) == BPF_EXIT) 298 continue; 299 300 if (insn->code == (BPF_JMP32 | BPF_JA)) { 301 if (i + 1 + insn->imm != tgt_idx) 302 continue; 303 if (check_add_overflow(insn->imm, delta, &imm)) 304 return -ERANGE; 305 insn->imm = imm; 306 } else { 307 if (i + 1 + insn->off != tgt_idx) 308 continue; 309 if (check_add_overflow(insn->off, delta, &off)) 310 return -ERANGE; 311 insn->off = off; 312 } 313 } 314 return 0; 315 } 316 317 static int adjust_subprog_starts_after_remove(struct bpf_verifier_env *env, 318 u32 off, u32 cnt) 319 { 320 int i, j; 321 322 /* find first prog starting at or after off (first to remove) */ 323 for (i = 0; i < env->subprog_cnt; i++) 324 if (env->subprog_info[i].start >= off) 325 break; 326 /* find first prog starting at or after off + cnt (first to stay) */ 327 for (j = i; j < env->subprog_cnt; j++) 328 if (env->subprog_info[j].start >= off + cnt) 329 break; 330 /* if j doesn't start exactly at off + cnt, we are just removing 331 * the front of previous prog 332 */ 333 if (env->subprog_info[j].start != off + cnt) 334 j--; 335 336 if (j > i) { 337 struct bpf_prog_aux *aux = env->prog->aux; 338 int move; 339 340 /* move fake 'exit' subprog as well */ 341 move = env->subprog_cnt + 1 - j; 342 343 memmove(env->subprog_info + i, 344 env->subprog_info + j, 345 sizeof(*env->subprog_info) * move); 346 env->subprog_cnt -= j - i; 347 348 /* remove func_info */ 349 if (aux->func_info) { 350 move = aux->func_info_cnt - j; 351 352 memmove(aux->func_info + i, 353 aux->func_info + j, 354 sizeof(*aux->func_info) * move); 355 aux->func_info_cnt -= j - i; 356 /* func_info->insn_off is set after all code rewrites, 357 * in adjust_btf_func() - no need to adjust 358 */ 359 } 360 } else { 361 /* convert i from "first prog to remove" to "first to adjust" */ 362 if (env->subprog_info[i].start == off) 363 i++; 364 } 365 366 /* update fake 'exit' subprog as well */ 367 for (; i <= env->subprog_cnt; i++) 368 env->subprog_info[i].start -= cnt; 369 370 return 0; 371 } 372 373 static int bpf_adj_linfo_after_remove(struct bpf_verifier_env *env, u32 off, 374 u32 cnt) 375 { 376 struct bpf_prog *prog = env->prog; 377 u32 i, l_off, l_cnt, nr_linfo; 378 struct bpf_line_info *linfo; 379 380 nr_linfo = prog->aux->nr_linfo; 381 if (!nr_linfo) 382 return 0; 383 384 linfo = prog->aux->linfo; 385 386 /* find first line info to remove, count lines to be removed */ 387 for (i = 0; i < nr_linfo; i++) 388 if (linfo[i].insn_off >= off) 389 break; 390 391 l_off = i; 392 l_cnt = 0; 393 for (; i < nr_linfo; i++) 394 if (linfo[i].insn_off < off + cnt) 395 l_cnt++; 396 else 397 break; 398 399 /* First live insn doesn't match first live linfo, it needs to "inherit" 400 * last removed linfo. prog is already modified, so prog->len == off 401 * means no live instructions after (tail of the program was removed). 402 */ 403 if (prog->len != off && l_cnt && 404 (i == nr_linfo || linfo[i].insn_off != off + cnt)) { 405 l_cnt--; 406 linfo[--i].insn_off = off + cnt; 407 } 408 409 /* remove the line info which refer to the removed instructions */ 410 if (l_cnt) { 411 memmove(linfo + l_off, linfo + i, 412 sizeof(*linfo) * (nr_linfo - i)); 413 414 prog->aux->nr_linfo -= l_cnt; 415 nr_linfo = prog->aux->nr_linfo; 416 } 417 418 /* pull all linfo[i].insn_off >= off + cnt in by cnt */ 419 for (i = l_off; i < nr_linfo; i++) 420 linfo[i].insn_off -= cnt; 421 422 /* fix up all subprogs (incl. 'exit') which start >= off */ 423 for (i = 0; i <= env->subprog_cnt; i++) 424 if (env->subprog_info[i].linfo_idx > l_off) { 425 /* program may have started in the removed region but 426 * may not be fully removed 427 */ 428 if (env->subprog_info[i].linfo_idx >= l_off + l_cnt) 429 env->subprog_info[i].linfo_idx -= l_cnt; 430 else 431 env->subprog_info[i].linfo_idx = l_off; 432 } 433 434 return 0; 435 } 436 437 /* 438 * Clean up dynamically allocated fields of aux data for instructions [start, ...] 439 */ 440 void bpf_clear_insn_aux_data(struct bpf_verifier_env *env, int start, int len) 441 { 442 struct bpf_insn_aux_data *aux_data = env->insn_aux_data; 443 struct bpf_insn *insns = env->prog->insnsi; 444 int end = start + len; 445 int i; 446 447 for (i = start; i < end; i++) { 448 if (aux_data[i].jt) { 449 kvfree(aux_data[i].jt); 450 aux_data[i].jt = NULL; 451 } 452 453 if (bpf_is_ldimm64(&insns[i])) 454 i++; 455 } 456 } 457 458 static int verifier_remove_insns(struct bpf_verifier_env *env, u32 off, u32 cnt) 459 { 460 struct bpf_insn_aux_data *aux_data = env->insn_aux_data; 461 unsigned int orig_prog_len = env->prog->len; 462 int err; 463 464 if (bpf_prog_is_offloaded(env->prog->aux)) 465 bpf_prog_offload_remove_insns(env, off, cnt); 466 467 /* Should be called before bpf_remove_insns, as it uses prog->insnsi */ 468 bpf_clear_insn_aux_data(env, off, cnt); 469 470 err = bpf_remove_insns(env->prog, off, cnt); 471 if (err) 472 return err; 473 474 err = adjust_subprog_starts_after_remove(env, off, cnt); 475 if (err) 476 return err; 477 478 err = bpf_adj_linfo_after_remove(env, off, cnt); 479 if (err) 480 return err; 481 482 adjust_insn_arrays_after_remove(env, off, cnt); 483 484 memmove(aux_data + off, aux_data + off + cnt, 485 sizeof(*aux_data) * (orig_prog_len - off - cnt)); 486 487 return 0; 488 } 489 490 static const struct bpf_insn NOP = BPF_JMP_IMM(BPF_JA, 0, 0, 0); 491 static const struct bpf_insn MAY_GOTO_0 = BPF_RAW_INSN(BPF_JMP | BPF_JCOND, 0, 0, 0, 0); 492 493 bool bpf_insn_is_cond_jump(u8 code) 494 { 495 u8 op; 496 497 op = BPF_OP(code); 498 if (BPF_CLASS(code) == BPF_JMP32) 499 return op != BPF_JA; 500 501 if (BPF_CLASS(code) != BPF_JMP) 502 return false; 503 504 return op != BPF_JA && op != BPF_EXIT && op != BPF_CALL; 505 } 506 507 void bpf_opt_hard_wire_dead_code_branches(struct bpf_verifier_env *env) 508 { 509 struct bpf_insn_aux_data *aux_data = env->insn_aux_data; 510 struct bpf_insn ja = BPF_JMP_IMM(BPF_JA, 0, 0, 0); 511 struct bpf_insn *insn = env->prog->insnsi; 512 const int insn_cnt = env->prog->len; 513 int i; 514 515 for (i = 0; i < insn_cnt; i++, insn++) { 516 if (!bpf_insn_is_cond_jump(insn->code)) 517 continue; 518 519 if (!aux_data[i + 1].seen) 520 ja.off = insn->off; 521 else if (!aux_data[i + 1 + insn->off].seen) 522 ja.off = 0; 523 else 524 continue; 525 526 if (bpf_prog_is_offloaded(env->prog->aux)) 527 bpf_prog_offload_replace_insn(env, i, &ja); 528 529 memcpy(insn, &ja, sizeof(ja)); 530 } 531 } 532 533 int bpf_opt_remove_dead_code(struct bpf_verifier_env *env) 534 { 535 struct bpf_insn_aux_data *aux_data = env->insn_aux_data; 536 int insn_cnt = env->prog->len; 537 int i, err; 538 539 for (i = 0; i < insn_cnt; i++) { 540 int j; 541 542 j = 0; 543 while (i + j < insn_cnt && !aux_data[i + j].seen) 544 j++; 545 if (!j) 546 continue; 547 548 err = verifier_remove_insns(env, i, j); 549 if (err) 550 return err; 551 insn_cnt = env->prog->len; 552 } 553 554 return 0; 555 } 556 557 int bpf_opt_remove_nops(struct bpf_verifier_env *env) 558 { 559 struct bpf_insn *insn = env->prog->insnsi; 560 int insn_cnt = env->prog->len; 561 bool is_may_goto_0, is_ja; 562 int i, err; 563 564 for (i = 0; i < insn_cnt; i++) { 565 is_may_goto_0 = !memcmp(&insn[i], &MAY_GOTO_0, sizeof(MAY_GOTO_0)); 566 is_ja = !memcmp(&insn[i], &NOP, sizeof(NOP)); 567 568 if (!is_may_goto_0 && !is_ja) 569 continue; 570 571 err = verifier_remove_insns(env, i, 1); 572 if (err) 573 return err; 574 insn_cnt--; 575 /* Go back one insn to catch may_goto +1; may_goto +0 sequence */ 576 i -= (is_may_goto_0 && i > 0) ? 2 : 1; 577 } 578 579 return 0; 580 } 581 582 int bpf_opt_subreg_zext_lo32_rnd_hi32(struct bpf_verifier_env *env, 583 const union bpf_attr *attr) 584 { 585 struct bpf_insn *patch; 586 /* use env->insn_buf as two independent buffers */ 587 struct bpf_insn *zext_patch = env->insn_buf; 588 struct bpf_insn *rnd_hi32_patch = &env->insn_buf[2]; 589 struct bpf_insn_aux_data *aux = env->insn_aux_data; 590 int i, patch_len, delta = 0, len = env->prog->len; 591 struct bpf_insn *insns = env->prog->insnsi; 592 struct bpf_prog *new_prog; 593 bool rnd_hi32; 594 595 rnd_hi32 = attr->prog_flags & BPF_F_TEST_RND_HI32; 596 zext_patch[1] = BPF_ZEXT_REG(0); 597 rnd_hi32_patch[1] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_AX, 0); 598 rnd_hi32_patch[2] = BPF_ALU64_IMM(BPF_LSH, BPF_REG_AX, 32); 599 rnd_hi32_patch[3] = BPF_ALU64_REG(BPF_OR, 0, BPF_REG_AX); 600 for (i = 0; i < len; i++) { 601 int adj_idx = i + delta; 602 struct bpf_insn insn; 603 int load_reg; 604 605 insn = insns[adj_idx]; 606 load_reg = insn_def_regno(&insn); 607 if (!aux[adj_idx].zext_dst) { 608 u8 code, class; 609 u32 imm_rnd; 610 611 if (!rnd_hi32) 612 continue; 613 614 code = insn.code; 615 class = BPF_CLASS(code); 616 if (load_reg == -1) 617 continue; 618 619 /* NOTE: arg "reg" (the fourth one) is only used for 620 * BPF_STX + SRC_OP, so it is safe to pass NULL 621 * here. 622 */ 623 if (bpf_is_reg64(&insn, load_reg, NULL, DST_OP)) { 624 if (class == BPF_LD && 625 BPF_MODE(code) == BPF_IMM) 626 i++; 627 continue; 628 } 629 630 /* ctx load could be transformed into wider load. */ 631 if (class == BPF_LDX && 632 aux[adj_idx].ptr_type == PTR_TO_CTX) 633 continue; 634 635 imm_rnd = get_random_u32(); 636 rnd_hi32_patch[0] = insn; 637 rnd_hi32_patch[1].imm = imm_rnd; 638 rnd_hi32_patch[3].dst_reg = load_reg; 639 patch = rnd_hi32_patch; 640 patch_len = 4; 641 goto apply_patch_buffer; 642 } 643 644 /* Add in an zero-extend instruction if a) the JIT has requested 645 * it or b) it's a CMPXCHG. 646 * 647 * The latter is because: BPF_CMPXCHG always loads a value into 648 * R0, therefore always zero-extends. However some archs' 649 * equivalent instruction only does this load when the 650 * comparison is successful. This detail of CMPXCHG is 651 * orthogonal to the general zero-extension behaviour of the 652 * CPU, so it's treated independently of bpf_jit_needs_zext. 653 */ 654 if (!bpf_jit_needs_zext() && !is_cmpxchg_insn(&insn)) 655 continue; 656 657 /* Zero-extension is done by the caller. */ 658 if (bpf_pseudo_kfunc_call(&insn)) 659 continue; 660 661 if (verifier_bug_if(load_reg == -1, env, 662 "zext_dst is set, but no reg is defined")) 663 return -EFAULT; 664 665 zext_patch[0] = insn; 666 zext_patch[1].dst_reg = load_reg; 667 zext_patch[1].src_reg = load_reg; 668 patch = zext_patch; 669 patch_len = 2; 670 apply_patch_buffer: 671 new_prog = bpf_patch_insn_data(env, adj_idx, patch, patch_len); 672 if (!new_prog) 673 return -ENOMEM; 674 env->prog = new_prog; 675 insns = new_prog->insnsi; 676 aux = env->insn_aux_data; 677 delta += patch_len - 1; 678 } 679 680 return 0; 681 } 682 683 /* convert load instructions that access fields of a context type into a 684 * sequence of instructions that access fields of the underlying structure: 685 * struct __sk_buff -> struct sk_buff 686 * struct bpf_sock_ops -> struct sock 687 */ 688 int bpf_convert_ctx_accesses(struct bpf_verifier_env *env) 689 { 690 struct bpf_subprog_info *subprogs = env->subprog_info; 691 const struct bpf_verifier_ops *ops = env->ops; 692 int i, cnt, size, ctx_field_size, ret, delta = 0, epilogue_cnt = 0; 693 const int insn_cnt = env->prog->len; 694 struct bpf_insn *epilogue_buf = env->epilogue_buf; 695 struct bpf_insn *insn_buf = env->insn_buf; 696 struct bpf_insn *insn; 697 u32 target_size, size_default, off; 698 struct bpf_prog *new_prog; 699 enum bpf_access_type type; 700 bool is_narrower_load; 701 int epilogue_idx = 0; 702 703 if (ops->gen_epilogue) { 704 epilogue_cnt = ops->gen_epilogue(epilogue_buf, env->prog, 705 -(subprogs[0].stack_depth + 8)); 706 if (epilogue_cnt >= INSN_BUF_SIZE) { 707 verifier_bug(env, "epilogue is too long"); 708 return -EFAULT; 709 } else if (epilogue_cnt) { 710 /* Save the ARG_PTR_TO_CTX for the epilogue to use */ 711 cnt = 0; 712 subprogs[0].stack_depth += 8; 713 insn_buf[cnt++] = BPF_STX_MEM(BPF_DW, BPF_REG_FP, BPF_REG_1, 714 -subprogs[0].stack_depth); 715 insn_buf[cnt++] = env->prog->insnsi[0]; 716 new_prog = bpf_patch_insn_data(env, 0, insn_buf, cnt); 717 if (!new_prog) 718 return -ENOMEM; 719 env->prog = new_prog; 720 delta += cnt - 1; 721 722 ret = add_kfunc_in_insns(env, epilogue_buf, epilogue_cnt - 1); 723 if (ret < 0) 724 return ret; 725 } 726 } 727 728 if (ops->gen_prologue || env->seen_direct_write) { 729 if (!ops->gen_prologue) { 730 verifier_bug(env, "gen_prologue is null"); 731 return -EFAULT; 732 } 733 cnt = ops->gen_prologue(insn_buf, env->seen_direct_write, 734 env->prog); 735 if (cnt >= INSN_BUF_SIZE) { 736 verifier_bug(env, "prologue is too long"); 737 return -EFAULT; 738 } else if (cnt) { 739 new_prog = bpf_patch_insn_data(env, 0, insn_buf, cnt); 740 if (!new_prog) 741 return -ENOMEM; 742 743 env->prog = new_prog; 744 delta += cnt - 1; 745 746 ret = add_kfunc_in_insns(env, insn_buf, cnt - 1); 747 if (ret < 0) 748 return ret; 749 } 750 } 751 752 if (delta) 753 WARN_ON(adjust_jmp_off(env->prog, 0, delta)); 754 755 if (bpf_prog_is_offloaded(env->prog->aux)) 756 return 0; 757 758 insn = env->prog->insnsi + delta; 759 760 for (i = 0; i < insn_cnt; i++, insn++) { 761 bpf_convert_ctx_access_t convert_ctx_access; 762 u8 mode; 763 764 if (env->insn_aux_data[i + delta].nospec) { 765 WARN_ON_ONCE(env->insn_aux_data[i + delta].alu_state); 766 struct bpf_insn *patch = insn_buf; 767 768 *patch++ = BPF_ST_NOSPEC(); 769 *patch++ = *insn; 770 cnt = patch - insn_buf; 771 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 772 if (!new_prog) 773 return -ENOMEM; 774 775 delta += cnt - 1; 776 env->prog = new_prog; 777 insn = new_prog->insnsi + i + delta; 778 /* This can not be easily merged with the 779 * nospec_result-case, because an insn may require a 780 * nospec before and after itself. Therefore also do not 781 * 'continue' here but potentially apply further 782 * patching to insn. *insn should equal patch[1] now. 783 */ 784 } 785 786 if (insn->code == (BPF_LDX | BPF_MEM | BPF_B) || 787 insn->code == (BPF_LDX | BPF_MEM | BPF_H) || 788 insn->code == (BPF_LDX | BPF_MEM | BPF_W) || 789 insn->code == (BPF_LDX | BPF_MEM | BPF_DW) || 790 insn->code == (BPF_LDX | BPF_MEMSX | BPF_B) || 791 insn->code == (BPF_LDX | BPF_MEMSX | BPF_H) || 792 insn->code == (BPF_LDX | BPF_MEMSX | BPF_W)) { 793 type = BPF_READ; 794 } else if (insn->code == (BPF_STX | BPF_MEM | BPF_B) || 795 insn->code == (BPF_STX | BPF_MEM | BPF_H) || 796 insn->code == (BPF_STX | BPF_MEM | BPF_W) || 797 insn->code == (BPF_STX | BPF_MEM | BPF_DW) || 798 insn->code == (BPF_ST | BPF_MEM | BPF_B) || 799 insn->code == (BPF_ST | BPF_MEM | BPF_H) || 800 insn->code == (BPF_ST | BPF_MEM | BPF_W) || 801 insn->code == (BPF_ST | BPF_MEM | BPF_DW)) { 802 type = BPF_WRITE; 803 } else if ((insn->code == (BPF_STX | BPF_ATOMIC | BPF_B) || 804 insn->code == (BPF_STX | BPF_ATOMIC | BPF_H) || 805 insn->code == (BPF_STX | BPF_ATOMIC | BPF_W) || 806 insn->code == (BPF_STX | BPF_ATOMIC | BPF_DW)) && 807 env->insn_aux_data[i + delta].ptr_type == PTR_TO_ARENA) { 808 insn->code = BPF_STX | BPF_PROBE_ATOMIC | BPF_SIZE(insn->code); 809 env->prog->aux->num_exentries++; 810 continue; 811 } else if (insn->code == (BPF_JMP | BPF_EXIT) && 812 epilogue_cnt && 813 i + delta < subprogs[1].start) { 814 /* Generate epilogue for the main prog */ 815 if (epilogue_idx) { 816 /* jump back to the earlier generated epilogue */ 817 insn_buf[0] = BPF_JMP32_A(epilogue_idx - i - delta - 1); 818 cnt = 1; 819 } else { 820 memcpy(insn_buf, epilogue_buf, 821 epilogue_cnt * sizeof(*epilogue_buf)); 822 cnt = epilogue_cnt; 823 /* epilogue_idx cannot be 0. It must have at 824 * least one ctx ptr saving insn before the 825 * epilogue. 826 */ 827 epilogue_idx = i + delta; 828 } 829 goto patch_insn_buf; 830 } else { 831 continue; 832 } 833 834 if (type == BPF_WRITE && 835 env->insn_aux_data[i + delta].nospec_result) { 836 /* nospec_result is only used to mitigate Spectre v4 and 837 * to limit verification-time for Spectre v1. 838 */ 839 struct bpf_insn *patch = insn_buf; 840 841 *patch++ = *insn; 842 *patch++ = BPF_ST_NOSPEC(); 843 cnt = patch - insn_buf; 844 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 845 if (!new_prog) 846 return -ENOMEM; 847 848 delta += cnt - 1; 849 env->prog = new_prog; 850 insn = new_prog->insnsi + i + delta; 851 continue; 852 } 853 854 switch ((int)env->insn_aux_data[i + delta].ptr_type) { 855 case PTR_TO_CTX: 856 if (!ops->convert_ctx_access) 857 continue; 858 convert_ctx_access = ops->convert_ctx_access; 859 break; 860 case PTR_TO_SOCKET: 861 case PTR_TO_SOCK_COMMON: 862 convert_ctx_access = bpf_sock_convert_ctx_access; 863 break; 864 case PTR_TO_TCP_SOCK: 865 convert_ctx_access = bpf_tcp_sock_convert_ctx_access; 866 break; 867 case PTR_TO_XDP_SOCK: 868 convert_ctx_access = bpf_xdp_sock_convert_ctx_access; 869 break; 870 case PTR_TO_BTF_ID: 871 case PTR_TO_BTF_ID | PTR_UNTRUSTED: 872 /* PTR_TO_BTF_ID | MEM_ALLOC always has a valid lifetime, unlike 873 * PTR_TO_BTF_ID, and an active ref_obj_id, but the same cannot 874 * be said once it is marked PTR_UNTRUSTED, hence we must handle 875 * any faults for loads into such types. BPF_WRITE is disallowed 876 * for this case. 877 */ 878 case PTR_TO_BTF_ID | MEM_ALLOC | PTR_UNTRUSTED: 879 case PTR_TO_MEM | MEM_RDONLY | PTR_UNTRUSTED: 880 if (type == BPF_READ) { 881 if (BPF_MODE(insn->code) == BPF_MEM) 882 insn->code = BPF_LDX | BPF_PROBE_MEM | 883 BPF_SIZE((insn)->code); 884 else 885 insn->code = BPF_LDX | BPF_PROBE_MEMSX | 886 BPF_SIZE((insn)->code); 887 env->prog->aux->num_exentries++; 888 } 889 continue; 890 case PTR_TO_ARENA: 891 if (BPF_MODE(insn->code) == BPF_MEMSX) { 892 if (!bpf_jit_supports_insn(insn, true)) { 893 verbose(env, "sign extending loads from arena are not supported yet\n"); 894 return -EOPNOTSUPP; 895 } 896 insn->code = BPF_CLASS(insn->code) | BPF_PROBE_MEM32SX | BPF_SIZE(insn->code); 897 } else { 898 insn->code = BPF_CLASS(insn->code) | BPF_PROBE_MEM32 | BPF_SIZE(insn->code); 899 } 900 env->prog->aux->num_exentries++; 901 continue; 902 default: 903 continue; 904 } 905 906 ctx_field_size = env->insn_aux_data[i + delta].ctx_field_size; 907 size = BPF_LDST_BYTES(insn); 908 mode = BPF_MODE(insn->code); 909 910 /* If the read access is a narrower load of the field, 911 * convert to a 4/8-byte load, to minimum program type specific 912 * convert_ctx_access changes. If conversion is successful, 913 * we will apply proper mask to the result. 914 */ 915 is_narrower_load = size < ctx_field_size; 916 size_default = bpf_ctx_off_adjust_machine(ctx_field_size); 917 off = insn->off; 918 if (is_narrower_load) { 919 u8 size_code; 920 921 if (type == BPF_WRITE) { 922 verifier_bug(env, "narrow ctx access misconfigured"); 923 return -EFAULT; 924 } 925 926 size_code = BPF_H; 927 if (ctx_field_size == 4) 928 size_code = BPF_W; 929 else if (ctx_field_size == 8) 930 size_code = BPF_DW; 931 932 insn->off = off & ~(size_default - 1); 933 insn->code = BPF_LDX | BPF_MEM | size_code; 934 } 935 936 target_size = 0; 937 cnt = convert_ctx_access(type, insn, insn_buf, env->prog, 938 &target_size); 939 if (cnt == 0 || cnt >= INSN_BUF_SIZE || 940 (ctx_field_size && !target_size)) { 941 verifier_bug(env, "error during ctx access conversion (%d)", cnt); 942 return -EFAULT; 943 } 944 945 if (is_narrower_load && size < target_size) { 946 u8 shift = bpf_ctx_narrow_access_offset( 947 off, size, size_default) * 8; 948 if (shift && cnt + 1 >= INSN_BUF_SIZE) { 949 verifier_bug(env, "narrow ctx load misconfigured"); 950 return -EFAULT; 951 } 952 if (ctx_field_size <= 4) { 953 if (shift) 954 insn_buf[cnt++] = BPF_ALU32_IMM(BPF_RSH, 955 insn->dst_reg, 956 shift); 957 insn_buf[cnt++] = BPF_ALU32_IMM(BPF_AND, insn->dst_reg, 958 (1 << size * 8) - 1); 959 } else { 960 if (shift) 961 insn_buf[cnt++] = BPF_ALU64_IMM(BPF_RSH, 962 insn->dst_reg, 963 shift); 964 insn_buf[cnt++] = BPF_ALU32_IMM(BPF_AND, insn->dst_reg, 965 (1ULL << size * 8) - 1); 966 } 967 } 968 if (mode == BPF_MEMSX) 969 insn_buf[cnt++] = BPF_RAW_INSN(BPF_ALU64 | BPF_MOV | BPF_X, 970 insn->dst_reg, insn->dst_reg, 971 size * 8, 0); 972 973 patch_insn_buf: 974 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 975 if (!new_prog) 976 return -ENOMEM; 977 978 delta += cnt - 1; 979 980 /* keep walking new program and skip insns we just inserted */ 981 env->prog = new_prog; 982 insn = new_prog->insnsi + i + delta; 983 } 984 985 return 0; 986 } 987 988 static u32 *bpf_dup_subprog_starts(struct bpf_verifier_env *env) 989 { 990 u32 *starts = NULL; 991 992 starts = kvmalloc_objs(u32, env->subprog_cnt, GFP_KERNEL_ACCOUNT); 993 if (starts) { 994 for (int i = 0; i < env->subprog_cnt; i++) 995 starts[i] = env->subprog_info[i].start; 996 } 997 return starts; 998 } 999 1000 static void bpf_restore_subprog_starts(struct bpf_verifier_env *env, u32 *orig_starts) 1001 { 1002 for (int i = 0; i < env->subprog_cnt; i++) 1003 env->subprog_info[i].start = orig_starts[i]; 1004 /* restore the start of fake 'exit' subprog as well */ 1005 env->subprog_info[env->subprog_cnt].start = env->prog->len; 1006 } 1007 1008 struct bpf_insn_aux_data *bpf_dup_insn_aux_data(struct bpf_verifier_env *env) 1009 { 1010 size_t size; 1011 void *new_aux; 1012 1013 size = array_size(sizeof(struct bpf_insn_aux_data), env->prog->len); 1014 new_aux = __vmalloc(size, GFP_KERNEL_ACCOUNT); 1015 if (new_aux) 1016 memcpy(new_aux, env->insn_aux_data, size); 1017 return new_aux; 1018 } 1019 1020 void bpf_restore_insn_aux_data(struct bpf_verifier_env *env, 1021 struct bpf_insn_aux_data *orig_insn_aux) 1022 { 1023 /* the expanded elements are zero-filled, so no special handling is required */ 1024 vfree(env->insn_aux_data); 1025 env->insn_aux_data = orig_insn_aux; 1026 } 1027 1028 static int jit_subprogs(struct bpf_verifier_env *env) 1029 { 1030 struct bpf_prog *prog = env->prog, **func, *tmp; 1031 int i, j, subprog_start, subprog_end = 0, len, subprog; 1032 struct bpf_map *map_ptr; 1033 struct bpf_insn *insn; 1034 void *old_bpf_func; 1035 int err, num_exentries; 1036 1037 for (i = 0, insn = prog->insnsi; i < prog->len; i++, insn++) { 1038 if (!bpf_pseudo_func(insn) && !bpf_pseudo_call(insn)) 1039 continue; 1040 1041 /* Upon error here we cannot fall back to interpreter but 1042 * need a hard reject of the program. Thus -EFAULT is 1043 * propagated in any case. 1044 */ 1045 subprog = bpf_find_subprog(env, i + insn->imm + 1); 1046 if (verifier_bug_if(subprog < 0, env, "No program to jit at insn %d", 1047 i + insn->imm + 1)) 1048 return -EFAULT; 1049 /* temporarily remember subprog id inside insn instead of 1050 * aux_data, since next loop will split up all insns into funcs 1051 */ 1052 insn->off = subprog; 1053 /* remember original imm in case JIT fails and fallback 1054 * to interpreter will be needed 1055 */ 1056 env->insn_aux_data[i].call_imm = insn->imm; 1057 /* point imm to __bpf_call_base+1 from JITs point of view */ 1058 insn->imm = 1; 1059 if (bpf_pseudo_func(insn)) { 1060 #if defined(MODULES_VADDR) 1061 u64 addr = MODULES_VADDR; 1062 #else 1063 u64 addr = VMALLOC_START; 1064 #endif 1065 /* jit (e.g. x86_64) may emit fewer instructions 1066 * if it learns a u32 imm is the same as a u64 imm. 1067 * Set close enough to possible prog address. 1068 */ 1069 insn[0].imm = (u32)addr; 1070 insn[1].imm = addr >> 32; 1071 } 1072 } 1073 1074 err = bpf_prog_alloc_jited_linfo(prog); 1075 if (err) 1076 goto out_undo_insn; 1077 1078 err = -ENOMEM; 1079 func = kzalloc_objs(prog, env->subprog_cnt); 1080 if (!func) 1081 goto out_undo_insn; 1082 1083 for (i = 0; i < env->subprog_cnt; i++) { 1084 subprog_start = subprog_end; 1085 subprog_end = env->subprog_info[i + 1].start; 1086 1087 len = subprog_end - subprog_start; 1088 /* bpf_prog_run() doesn't call subprogs directly, 1089 * hence main prog stats include the runtime of subprogs. 1090 * subprogs don't have IDs and not reachable via prog_get_next_id 1091 * func[i]->stats will never be accessed and stays NULL 1092 */ 1093 func[i] = bpf_prog_alloc_no_stats(bpf_prog_size(len), GFP_USER); 1094 if (!func[i]) 1095 goto out_free; 1096 memcpy(func[i]->insnsi, &prog->insnsi[subprog_start], 1097 len * sizeof(struct bpf_insn)); 1098 func[i]->type = prog->type; 1099 func[i]->len = len; 1100 if (bpf_prog_calc_tag(func[i])) 1101 goto out_free; 1102 func[i]->is_func = 1; 1103 func[i]->sleepable = prog->sleepable; 1104 func[i]->blinded = prog->blinded; 1105 func[i]->aux->func_idx = i; 1106 /* Below members will be freed only at prog->aux */ 1107 func[i]->aux->btf = prog->aux->btf; 1108 func[i]->aux->subprog_start = subprog_start; 1109 func[i]->aux->func_info = prog->aux->func_info; 1110 func[i]->aux->func_info_cnt = prog->aux->func_info_cnt; 1111 func[i]->aux->poke_tab = prog->aux->poke_tab; 1112 func[i]->aux->size_poke_tab = prog->aux->size_poke_tab; 1113 func[i]->aux->main_prog_aux = prog->aux; 1114 1115 for (j = 0; j < prog->aux->size_poke_tab; j++) { 1116 struct bpf_jit_poke_descriptor *poke; 1117 1118 poke = &prog->aux->poke_tab[j]; 1119 if (poke->insn_idx < subprog_end && 1120 poke->insn_idx >= subprog_start) 1121 poke->aux = func[i]->aux; 1122 } 1123 1124 func[i]->aux->name[0] = 'F'; 1125 func[i]->aux->stack_depth = env->subprog_info[i].stack_depth; 1126 if (env->subprog_info[i].priv_stack_mode == PRIV_STACK_ADAPTIVE) 1127 func[i]->aux->jits_use_priv_stack = true; 1128 1129 func[i]->jit_requested = 1; 1130 func[i]->blinding_requested = prog->blinding_requested; 1131 func[i]->aux->kfunc_tab = prog->aux->kfunc_tab; 1132 func[i]->aux->kfunc_btf_tab = prog->aux->kfunc_btf_tab; 1133 func[i]->aux->linfo = prog->aux->linfo; 1134 func[i]->aux->nr_linfo = prog->aux->nr_linfo; 1135 func[i]->aux->jited_linfo = prog->aux->jited_linfo; 1136 func[i]->aux->linfo_idx = env->subprog_info[i].linfo_idx; 1137 func[i]->aux->arena = prog->aux->arena; 1138 func[i]->aux->used_maps = env->used_maps; 1139 func[i]->aux->used_map_cnt = env->used_map_cnt; 1140 num_exentries = 0; 1141 insn = func[i]->insnsi; 1142 for (j = 0; j < func[i]->len; j++, insn++) { 1143 if (BPF_CLASS(insn->code) == BPF_LDX && 1144 (BPF_MODE(insn->code) == BPF_PROBE_MEM || 1145 BPF_MODE(insn->code) == BPF_PROBE_MEM32 || 1146 BPF_MODE(insn->code) == BPF_PROBE_MEM32SX || 1147 BPF_MODE(insn->code) == BPF_PROBE_MEMSX)) 1148 num_exentries++; 1149 if ((BPF_CLASS(insn->code) == BPF_STX || 1150 BPF_CLASS(insn->code) == BPF_ST) && 1151 BPF_MODE(insn->code) == BPF_PROBE_MEM32) 1152 num_exentries++; 1153 if (BPF_CLASS(insn->code) == BPF_STX && 1154 BPF_MODE(insn->code) == BPF_PROBE_ATOMIC) 1155 num_exentries++; 1156 } 1157 func[i]->aux->num_exentries = num_exentries; 1158 func[i]->aux->tail_call_reachable = env->subprog_info[i].tail_call_reachable; 1159 func[i]->aux->exception_cb = env->subprog_info[i].is_exception_cb; 1160 func[i]->aux->changes_pkt_data = env->subprog_info[i].changes_pkt_data; 1161 func[i]->aux->might_sleep = env->subprog_info[i].might_sleep; 1162 func[i]->aux->token = prog->aux->token; 1163 if (!i) 1164 func[i]->aux->exception_boundary = env->seen_exception; 1165 func[i] = bpf_int_jit_compile(env, func[i]); 1166 if (!func[i]->jited) { 1167 err = -ENOTSUPP; 1168 goto out_free; 1169 } 1170 cond_resched(); 1171 } 1172 1173 /* at this point all bpf functions were successfully JITed 1174 * now populate all bpf_calls with correct addresses and 1175 * run last pass of JIT 1176 */ 1177 for (i = 0; i < env->subprog_cnt; i++) { 1178 insn = func[i]->insnsi; 1179 for (j = 0; j < func[i]->len; j++, insn++) { 1180 if (bpf_pseudo_func(insn)) { 1181 subprog = insn->off; 1182 insn[0].imm = (u32)(long)func[subprog]->bpf_func; 1183 insn[1].imm = ((u64)(long)func[subprog]->bpf_func) >> 32; 1184 continue; 1185 } 1186 if (!bpf_pseudo_call(insn)) 1187 continue; 1188 subprog = insn->off; 1189 insn->imm = BPF_CALL_IMM(func[subprog]->bpf_func); 1190 } 1191 1192 /* we use the aux data to keep a list of the start addresses 1193 * of the JITed images for each function in the program 1194 * 1195 * for some architectures, such as powerpc64, the imm field 1196 * might not be large enough to hold the offset of the start 1197 * address of the callee's JITed image from __bpf_call_base 1198 * 1199 * in such cases, we can lookup the start address of a callee 1200 * by using its subprog id, available from the off field of 1201 * the call instruction, as an index for this list 1202 */ 1203 func[i]->aux->func = func; 1204 func[i]->aux->func_cnt = env->subprog_cnt - env->hidden_subprog_cnt; 1205 func[i]->aux->real_func_cnt = env->subprog_cnt; 1206 } 1207 for (i = 0; i < env->subprog_cnt; i++) { 1208 old_bpf_func = func[i]->bpf_func; 1209 tmp = bpf_int_jit_compile(env, func[i]); 1210 if (tmp != func[i] || func[i]->bpf_func != old_bpf_func) { 1211 verbose(env, "JIT doesn't support bpf-to-bpf calls\n"); 1212 err = -ENOTSUPP; 1213 goto out_free; 1214 } 1215 cond_resched(); 1216 } 1217 1218 /* 1219 * Cleanup func[i]->aux fields which aren't required 1220 * or can become invalid in future 1221 */ 1222 for (i = 0; i < env->subprog_cnt; i++) { 1223 func[i]->aux->used_maps = NULL; 1224 func[i]->aux->used_map_cnt = 0; 1225 } 1226 1227 /* finally lock prog and jit images for all functions and 1228 * populate kallsysm. Begin at the first subprogram, since 1229 * bpf_prog_load will add the kallsyms for the main program. 1230 */ 1231 for (i = 1; i < env->subprog_cnt; i++) { 1232 err = bpf_prog_lock_ro(func[i]); 1233 if (err) 1234 goto out_free; 1235 } 1236 1237 for (i = 1; i < env->subprog_cnt; i++) 1238 bpf_prog_kallsyms_add(func[i]); 1239 1240 /* Last step: make now unused interpreter insns from main 1241 * prog consistent for later dump requests, so they can 1242 * later look the same as if they were interpreted only. 1243 */ 1244 for (i = 0, insn = prog->insnsi; i < prog->len; i++, insn++) { 1245 if (bpf_pseudo_func(insn)) { 1246 insn[0].imm = env->insn_aux_data[i].call_imm; 1247 insn[1].imm = insn->off; 1248 insn->off = 0; 1249 continue; 1250 } 1251 if (!bpf_pseudo_call(insn)) 1252 continue; 1253 insn->off = env->insn_aux_data[i].call_imm; 1254 subprog = bpf_find_subprog(env, i + insn->off + 1); 1255 insn->imm = subprog; 1256 } 1257 1258 prog->jited = 1; 1259 prog->bpf_func = func[0]->bpf_func; 1260 prog->jited_len = func[0]->jited_len; 1261 prog->aux->extable = func[0]->aux->extable; 1262 prog->aux->num_exentries = func[0]->aux->num_exentries; 1263 prog->aux->func = func; 1264 prog->aux->func_cnt = env->subprog_cnt - env->hidden_subprog_cnt; 1265 prog->aux->real_func_cnt = env->subprog_cnt; 1266 prog->aux->bpf_exception_cb = (void *)func[env->exception_callback_subprog]->bpf_func; 1267 prog->aux->exception_boundary = func[0]->aux->exception_boundary; 1268 bpf_prog_jit_attempt_done(prog); 1269 return 0; 1270 out_free: 1271 /* We failed JIT'ing, so at this point we need to unregister poke 1272 * descriptors from subprogs, so that kernel is not attempting to 1273 * patch it anymore as we're freeing the subprog JIT memory. 1274 */ 1275 for (i = 0; i < prog->aux->size_poke_tab; i++) { 1276 map_ptr = prog->aux->poke_tab[i].tail_call.map; 1277 map_ptr->ops->map_poke_untrack(map_ptr, prog->aux); 1278 } 1279 /* At this point we're guaranteed that poke descriptors are not 1280 * live anymore. We can just unlink its descriptor table as it's 1281 * released with the main prog. 1282 */ 1283 for (i = 0; i < env->subprog_cnt; i++) { 1284 if (!func[i]) 1285 continue; 1286 func[i]->aux->poke_tab = NULL; 1287 bpf_jit_free(func[i]); 1288 } 1289 kfree(func); 1290 out_undo_insn: 1291 bpf_prog_jit_attempt_done(prog); 1292 return err; 1293 } 1294 1295 int bpf_jit_subprogs(struct bpf_verifier_env *env) 1296 { 1297 int err, i; 1298 bool blinded = false; 1299 struct bpf_insn *insn; 1300 struct bpf_prog *prog, *orig_prog; 1301 struct bpf_insn_aux_data *orig_insn_aux; 1302 u32 *orig_subprog_starts; 1303 1304 if (env->subprog_cnt <= 1) 1305 return 0; 1306 1307 prog = orig_prog = env->prog; 1308 if (bpf_prog_need_blind(prog)) { 1309 orig_insn_aux = bpf_dup_insn_aux_data(env); 1310 if (!orig_insn_aux) { 1311 err = -ENOMEM; 1312 goto out_cleanup; 1313 } 1314 orig_subprog_starts = bpf_dup_subprog_starts(env); 1315 if (!orig_subprog_starts) { 1316 vfree(orig_insn_aux); 1317 err = -ENOMEM; 1318 goto out_cleanup; 1319 } 1320 prog = bpf_jit_blind_constants(env, prog); 1321 if (IS_ERR(prog)) { 1322 err = -ENOMEM; 1323 prog = orig_prog; 1324 goto out_restore; 1325 } 1326 blinded = true; 1327 } 1328 1329 err = jit_subprogs(env); 1330 if (err) 1331 goto out_jit_err; 1332 1333 if (blinded) { 1334 bpf_jit_prog_release_other(prog, orig_prog); 1335 kvfree(orig_subprog_starts); 1336 vfree(orig_insn_aux); 1337 } 1338 1339 return 0; 1340 1341 out_jit_err: 1342 if (blinded) { 1343 bpf_jit_prog_release_other(orig_prog, prog); 1344 /* roll back to the clean original prog */ 1345 prog = env->prog = orig_prog; 1346 goto out_restore; 1347 } else { 1348 if (err != -EFAULT) { 1349 /* 1350 * We will fall back to interpreter mode when err is not -EFAULT, before 1351 * that, insn->off and insn->imm should be restored to their original 1352 * values since they were modified by jit_subprogs. 1353 */ 1354 for (i = 0, insn = prog->insnsi; i < prog->len; i++, insn++) { 1355 if (!bpf_pseudo_call(insn)) 1356 continue; 1357 insn->off = 0; 1358 insn->imm = env->insn_aux_data[i].call_imm; 1359 } 1360 } 1361 goto out_cleanup; 1362 } 1363 1364 out_restore: 1365 bpf_restore_subprog_starts(env, orig_subprog_starts); 1366 bpf_restore_insn_aux_data(env, orig_insn_aux); 1367 kvfree(orig_subprog_starts); 1368 out_cleanup: 1369 /* cleanup main prog to be interpreted */ 1370 prog->jit_requested = 0; 1371 prog->blinding_requested = 0; 1372 return err; 1373 } 1374 1375 int bpf_fixup_call_args(struct bpf_verifier_env *env) 1376 { 1377 #ifndef CONFIG_BPF_JIT_ALWAYS_ON 1378 struct bpf_prog *prog = env->prog; 1379 struct bpf_insn *insn = prog->insnsi; 1380 bool has_kfunc_call = bpf_prog_has_kfunc_call(prog); 1381 int i, depth; 1382 #endif 1383 int err = 0; 1384 1385 if (env->prog->jit_requested && 1386 !bpf_prog_is_offloaded(env->prog->aux)) { 1387 err = bpf_jit_subprogs(env); 1388 if (err == 0) 1389 return 0; 1390 if (err == -EFAULT) 1391 return err; 1392 } 1393 #ifndef CONFIG_BPF_JIT_ALWAYS_ON 1394 if (has_kfunc_call) { 1395 verbose(env, "calling kernel functions are not allowed in non-JITed programs\n"); 1396 return -EINVAL; 1397 } 1398 if (env->subprog_cnt > 1 && env->prog->aux->tail_call_reachable) { 1399 /* When JIT fails the progs with bpf2bpf calls and tail_calls 1400 * have to be rejected, since interpreter doesn't support them yet. 1401 */ 1402 verbose(env, "tail_calls are not allowed in non-JITed programs with bpf-to-bpf calls\n"); 1403 return -EINVAL; 1404 } 1405 for (i = 0; i < prog->len; i++, insn++) { 1406 if (bpf_pseudo_func(insn)) { 1407 /* When JIT fails the progs with callback calls 1408 * have to be rejected, since interpreter doesn't support them yet. 1409 */ 1410 verbose(env, "callbacks are not allowed in non-JITed programs\n"); 1411 return -EINVAL; 1412 } 1413 1414 if (!bpf_pseudo_call(insn)) 1415 continue; 1416 depth = get_callee_stack_depth(env, insn, i); 1417 if (depth < 0) 1418 return depth; 1419 bpf_patch_call_args(insn, depth); 1420 } 1421 err = 0; 1422 #endif 1423 return err; 1424 } 1425 1426 1427 /* The function requires that first instruction in 'patch' is insnsi[prog->len - 1] */ 1428 static int add_hidden_subprog(struct bpf_verifier_env *env, struct bpf_insn *patch, int len) 1429 { 1430 struct bpf_subprog_info *info = env->subprog_info; 1431 int cnt = env->subprog_cnt; 1432 struct bpf_prog *prog; 1433 1434 /* We only reserve one slot for hidden subprogs in subprog_info. */ 1435 if (env->hidden_subprog_cnt) { 1436 verifier_bug(env, "only one hidden subprog supported"); 1437 return -EFAULT; 1438 } 1439 /* We're not patching any existing instruction, just appending the new 1440 * ones for the hidden subprog. Hence all of the adjustment operations 1441 * in bpf_patch_insn_data are no-ops. 1442 */ 1443 prog = bpf_patch_insn_data(env, env->prog->len - 1, patch, len); 1444 if (!prog) 1445 return -ENOMEM; 1446 env->prog = prog; 1447 info[cnt + 1].start = info[cnt].start; 1448 info[cnt].start = prog->len - len + 1; 1449 env->subprog_cnt++; 1450 env->hidden_subprog_cnt++; 1451 return 0; 1452 } 1453 1454 /* Do various post-verification rewrites in a single program pass. 1455 * These rewrites simplify JIT and interpreter implementations. 1456 */ 1457 int bpf_do_misc_fixups(struct bpf_verifier_env *env) 1458 { 1459 struct bpf_prog *prog = env->prog; 1460 enum bpf_attach_type eatype = prog->expected_attach_type; 1461 enum bpf_prog_type prog_type = resolve_prog_type(prog); 1462 struct bpf_insn *insn = prog->insnsi; 1463 const struct bpf_func_proto *fn; 1464 const int insn_cnt = prog->len; 1465 const struct bpf_map_ops *ops; 1466 struct bpf_insn_aux_data *aux; 1467 struct bpf_insn *insn_buf = env->insn_buf; 1468 struct bpf_prog *new_prog; 1469 struct bpf_map *map_ptr; 1470 int i, ret, cnt, delta = 0, cur_subprog = 0; 1471 struct bpf_subprog_info *subprogs = env->subprog_info; 1472 u16 stack_depth = subprogs[cur_subprog].stack_depth; 1473 u16 stack_depth_extra = 0; 1474 1475 if (env->seen_exception && !env->exception_callback_subprog) { 1476 struct bpf_insn *patch = insn_buf; 1477 1478 *patch++ = env->prog->insnsi[insn_cnt - 1]; 1479 *patch++ = BPF_MOV64_REG(BPF_REG_0, BPF_REG_1); 1480 *patch++ = BPF_EXIT_INSN(); 1481 ret = add_hidden_subprog(env, insn_buf, patch - insn_buf); 1482 if (ret < 0) 1483 return ret; 1484 prog = env->prog; 1485 insn = prog->insnsi; 1486 1487 env->exception_callback_subprog = env->subprog_cnt - 1; 1488 /* Don't update insn_cnt, as add_hidden_subprog always appends insns */ 1489 bpf_mark_subprog_exc_cb(env, env->exception_callback_subprog); 1490 } 1491 1492 for (i = 0; i < insn_cnt;) { 1493 if (insn->code == (BPF_ALU64 | BPF_MOV | BPF_X) && insn->imm) { 1494 if ((insn->off == BPF_ADDR_SPACE_CAST && insn->imm == 1) || 1495 (((struct bpf_map *)env->prog->aux->arena)->map_flags & BPF_F_NO_USER_CONV)) { 1496 /* convert to 32-bit mov that clears upper 32-bit */ 1497 insn->code = BPF_ALU | BPF_MOV | BPF_X; 1498 /* clear off and imm, so it's a normal 'wX = wY' from JIT pov */ 1499 insn->off = 0; 1500 insn->imm = 0; 1501 } /* cast from as(0) to as(1) should be handled by JIT */ 1502 goto next_insn; 1503 } 1504 1505 if (env->insn_aux_data[i + delta].needs_zext) 1506 /* Convert BPF_CLASS(insn->code) == BPF_ALU64 to 32-bit ALU */ 1507 insn->code = BPF_ALU | BPF_OP(insn->code) | BPF_SRC(insn->code); 1508 1509 /* Make sdiv/smod divide-by-minus-one exceptions impossible. */ 1510 if ((insn->code == (BPF_ALU64 | BPF_MOD | BPF_K) || 1511 insn->code == (BPF_ALU64 | BPF_DIV | BPF_K) || 1512 insn->code == (BPF_ALU | BPF_MOD | BPF_K) || 1513 insn->code == (BPF_ALU | BPF_DIV | BPF_K)) && 1514 insn->off == 1 && insn->imm == -1) { 1515 bool is64 = BPF_CLASS(insn->code) == BPF_ALU64; 1516 bool isdiv = BPF_OP(insn->code) == BPF_DIV; 1517 struct bpf_insn *patch = insn_buf; 1518 1519 if (isdiv) 1520 *patch++ = BPF_RAW_INSN((is64 ? BPF_ALU64 : BPF_ALU) | 1521 BPF_NEG | BPF_K, insn->dst_reg, 1522 0, 0, 0); 1523 else 1524 *patch++ = BPF_MOV32_IMM(insn->dst_reg, 0); 1525 1526 cnt = patch - insn_buf; 1527 1528 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1529 if (!new_prog) 1530 return -ENOMEM; 1531 1532 delta += cnt - 1; 1533 env->prog = prog = new_prog; 1534 insn = new_prog->insnsi + i + delta; 1535 goto next_insn; 1536 } 1537 1538 /* Make divide-by-zero and divide-by-minus-one exceptions impossible. */ 1539 if (insn->code == (BPF_ALU64 | BPF_MOD | BPF_X) || 1540 insn->code == (BPF_ALU64 | BPF_DIV | BPF_X) || 1541 insn->code == (BPF_ALU | BPF_MOD | BPF_X) || 1542 insn->code == (BPF_ALU | BPF_DIV | BPF_X)) { 1543 bool is64 = BPF_CLASS(insn->code) == BPF_ALU64; 1544 bool isdiv = BPF_OP(insn->code) == BPF_DIV; 1545 bool is_sdiv = isdiv && insn->off == 1; 1546 bool is_smod = !isdiv && insn->off == 1; 1547 struct bpf_insn *patch = insn_buf; 1548 1549 if (is_sdiv) { 1550 /* [R,W]x sdiv 0 -> 0 1551 * LLONG_MIN sdiv -1 -> LLONG_MIN 1552 * INT_MIN sdiv -1 -> INT_MIN 1553 */ 1554 *patch++ = BPF_MOV64_REG(BPF_REG_AX, insn->src_reg); 1555 *patch++ = BPF_RAW_INSN((is64 ? BPF_ALU64 : BPF_ALU) | 1556 BPF_ADD | BPF_K, BPF_REG_AX, 1557 0, 0, 1); 1558 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1559 BPF_JGT | BPF_K, BPF_REG_AX, 1560 0, 4, 1); 1561 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1562 BPF_JEQ | BPF_K, BPF_REG_AX, 1563 0, 1, 0); 1564 *patch++ = BPF_RAW_INSN((is64 ? BPF_ALU64 : BPF_ALU) | 1565 BPF_MOV | BPF_K, insn->dst_reg, 1566 0, 0, 0); 1567 /* BPF_NEG(LLONG_MIN) == -LLONG_MIN == LLONG_MIN */ 1568 *patch++ = BPF_RAW_INSN((is64 ? BPF_ALU64 : BPF_ALU) | 1569 BPF_NEG | BPF_K, insn->dst_reg, 1570 0, 0, 0); 1571 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1572 *patch++ = *insn; 1573 cnt = patch - insn_buf; 1574 } else if (is_smod) { 1575 /* [R,W]x mod 0 -> [R,W]x */ 1576 /* [R,W]x mod -1 -> 0 */ 1577 *patch++ = BPF_MOV64_REG(BPF_REG_AX, insn->src_reg); 1578 *patch++ = BPF_RAW_INSN((is64 ? BPF_ALU64 : BPF_ALU) | 1579 BPF_ADD | BPF_K, BPF_REG_AX, 1580 0, 0, 1); 1581 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1582 BPF_JGT | BPF_K, BPF_REG_AX, 1583 0, 3, 1); 1584 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1585 BPF_JEQ | BPF_K, BPF_REG_AX, 1586 0, 3 + (is64 ? 0 : 1), 1); 1587 *patch++ = BPF_MOV32_IMM(insn->dst_reg, 0); 1588 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1589 *patch++ = *insn; 1590 1591 if (!is64) { 1592 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1593 *patch++ = BPF_MOV32_REG(insn->dst_reg, insn->dst_reg); 1594 } 1595 cnt = patch - insn_buf; 1596 } else if (isdiv) { 1597 /* [R,W]x div 0 -> 0 */ 1598 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1599 BPF_JNE | BPF_K, insn->src_reg, 1600 0, 2, 0); 1601 *patch++ = BPF_ALU32_REG(BPF_XOR, insn->dst_reg, insn->dst_reg); 1602 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1603 *patch++ = *insn; 1604 cnt = patch - insn_buf; 1605 } else { 1606 /* [R,W]x mod 0 -> [R,W]x */ 1607 *patch++ = BPF_RAW_INSN((is64 ? BPF_JMP : BPF_JMP32) | 1608 BPF_JEQ | BPF_K, insn->src_reg, 1609 0, 1 + (is64 ? 0 : 1), 0); 1610 *patch++ = *insn; 1611 1612 if (!is64) { 1613 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1614 *patch++ = BPF_MOV32_REG(insn->dst_reg, insn->dst_reg); 1615 } 1616 cnt = patch - insn_buf; 1617 } 1618 1619 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1620 if (!new_prog) 1621 return -ENOMEM; 1622 1623 delta += cnt - 1; 1624 env->prog = prog = new_prog; 1625 insn = new_prog->insnsi + i + delta; 1626 goto next_insn; 1627 } 1628 1629 /* Make it impossible to de-reference a userspace address */ 1630 if (BPF_CLASS(insn->code) == BPF_LDX && 1631 (BPF_MODE(insn->code) == BPF_PROBE_MEM || 1632 BPF_MODE(insn->code) == BPF_PROBE_MEMSX)) { 1633 struct bpf_insn *patch = insn_buf; 1634 u64 uaddress_limit = bpf_arch_uaddress_limit(); 1635 1636 if (!uaddress_limit) 1637 goto next_insn; 1638 1639 *patch++ = BPF_MOV64_REG(BPF_REG_AX, insn->src_reg); 1640 if (insn->off) 1641 *patch++ = BPF_ALU64_IMM(BPF_ADD, BPF_REG_AX, insn->off); 1642 *patch++ = BPF_ALU64_IMM(BPF_RSH, BPF_REG_AX, 32); 1643 *patch++ = BPF_JMP_IMM(BPF_JLE, BPF_REG_AX, uaddress_limit >> 32, 2); 1644 *patch++ = *insn; 1645 *patch++ = BPF_JMP_IMM(BPF_JA, 0, 0, 1); 1646 *patch++ = BPF_MOV64_IMM(insn->dst_reg, 0); 1647 1648 cnt = patch - insn_buf; 1649 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1650 if (!new_prog) 1651 return -ENOMEM; 1652 1653 delta += cnt - 1; 1654 env->prog = prog = new_prog; 1655 insn = new_prog->insnsi + i + delta; 1656 goto next_insn; 1657 } 1658 1659 /* Implement LD_ABS and LD_IND with a rewrite, if supported by the program type. */ 1660 if (BPF_CLASS(insn->code) == BPF_LD && 1661 (BPF_MODE(insn->code) == BPF_ABS || 1662 BPF_MODE(insn->code) == BPF_IND)) { 1663 cnt = env->ops->gen_ld_abs(insn, insn_buf); 1664 if (cnt == 0 || cnt >= INSN_BUF_SIZE) { 1665 verifier_bug(env, "%d insns generated for ld_abs", cnt); 1666 return -EFAULT; 1667 } 1668 1669 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1670 if (!new_prog) 1671 return -ENOMEM; 1672 1673 delta += cnt - 1; 1674 env->prog = prog = new_prog; 1675 insn = new_prog->insnsi + i + delta; 1676 goto next_insn; 1677 } 1678 1679 /* Rewrite pointer arithmetic to mitigate speculation attacks. */ 1680 if (insn->code == (BPF_ALU64 | BPF_ADD | BPF_X) || 1681 insn->code == (BPF_ALU64 | BPF_SUB | BPF_X)) { 1682 const u8 code_add = BPF_ALU64 | BPF_ADD | BPF_X; 1683 const u8 code_sub = BPF_ALU64 | BPF_SUB | BPF_X; 1684 struct bpf_insn *patch = insn_buf; 1685 bool issrc, isneg, isimm; 1686 u32 off_reg; 1687 1688 aux = &env->insn_aux_data[i + delta]; 1689 if (!aux->alu_state || 1690 aux->alu_state == BPF_ALU_NON_POINTER) 1691 goto next_insn; 1692 1693 isneg = aux->alu_state & BPF_ALU_NEG_VALUE; 1694 issrc = (aux->alu_state & BPF_ALU_SANITIZE) == 1695 BPF_ALU_SANITIZE_SRC; 1696 isimm = aux->alu_state & BPF_ALU_IMMEDIATE; 1697 1698 off_reg = issrc ? insn->src_reg : insn->dst_reg; 1699 if (isimm) { 1700 *patch++ = BPF_MOV32_IMM(BPF_REG_AX, aux->alu_limit); 1701 } else { 1702 if (isneg) 1703 *patch++ = BPF_ALU64_IMM(BPF_MUL, off_reg, -1); 1704 *patch++ = BPF_MOV32_IMM(BPF_REG_AX, aux->alu_limit); 1705 *patch++ = BPF_ALU64_REG(BPF_SUB, BPF_REG_AX, off_reg); 1706 *patch++ = BPF_ALU64_REG(BPF_OR, BPF_REG_AX, off_reg); 1707 *patch++ = BPF_ALU64_IMM(BPF_NEG, BPF_REG_AX, 0); 1708 *patch++ = BPF_ALU64_IMM(BPF_ARSH, BPF_REG_AX, 63); 1709 *patch++ = BPF_ALU64_REG(BPF_AND, BPF_REG_AX, off_reg); 1710 } 1711 if (!issrc) 1712 *patch++ = BPF_MOV64_REG(insn->dst_reg, insn->src_reg); 1713 insn->src_reg = BPF_REG_AX; 1714 if (isneg) 1715 insn->code = insn->code == code_add ? 1716 code_sub : code_add; 1717 *patch++ = *insn; 1718 if (issrc && isneg && !isimm) 1719 *patch++ = BPF_ALU64_IMM(BPF_MUL, off_reg, -1); 1720 cnt = patch - insn_buf; 1721 1722 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1723 if (!new_prog) 1724 return -ENOMEM; 1725 1726 delta += cnt - 1; 1727 env->prog = prog = new_prog; 1728 insn = new_prog->insnsi + i + delta; 1729 goto next_insn; 1730 } 1731 1732 if (bpf_is_may_goto_insn(insn) && bpf_jit_supports_timed_may_goto()) { 1733 int stack_off_cnt = -stack_depth - 16; 1734 1735 /* 1736 * Two 8 byte slots, depth-16 stores the count, and 1737 * depth-8 stores the start timestamp of the loop. 1738 * 1739 * The starting value of count is BPF_MAX_TIMED_LOOPS 1740 * (0xffff). Every iteration loads it and subs it by 1, 1741 * until the value becomes 0 in AX (thus, 1 in stack), 1742 * after which we call arch_bpf_timed_may_goto, which 1743 * either sets AX to 0xffff to keep looping, or to 0 1744 * upon timeout. AX is then stored into the stack. In 1745 * the next iteration, we either see 0 and break out, or 1746 * continue iterating until the next time value is 0 1747 * after subtraction, rinse and repeat. 1748 */ 1749 stack_depth_extra = 16; 1750 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_AX, BPF_REG_10, stack_off_cnt); 1751 if (insn->off >= 0) 1752 insn_buf[1] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_AX, 0, insn->off + 5); 1753 else 1754 insn_buf[1] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_AX, 0, insn->off - 1); 1755 insn_buf[2] = BPF_ALU64_IMM(BPF_SUB, BPF_REG_AX, 1); 1756 insn_buf[3] = BPF_JMP_IMM(BPF_JNE, BPF_REG_AX, 0, 2); 1757 /* 1758 * AX is used as an argument to pass in stack_off_cnt 1759 * (to add to r10/fp), and also as the return value of 1760 * the call to arch_bpf_timed_may_goto. 1761 */ 1762 insn_buf[4] = BPF_MOV64_IMM(BPF_REG_AX, stack_off_cnt); 1763 insn_buf[5] = BPF_EMIT_CALL(arch_bpf_timed_may_goto); 1764 insn_buf[6] = BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_AX, stack_off_cnt); 1765 cnt = 7; 1766 1767 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1768 if (!new_prog) 1769 return -ENOMEM; 1770 1771 delta += cnt - 1; 1772 env->prog = prog = new_prog; 1773 insn = new_prog->insnsi + i + delta; 1774 goto next_insn; 1775 } else if (bpf_is_may_goto_insn(insn)) { 1776 int stack_off = -stack_depth - 8; 1777 1778 stack_depth_extra = 8; 1779 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_AX, BPF_REG_10, stack_off); 1780 if (insn->off >= 0) 1781 insn_buf[1] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_AX, 0, insn->off + 2); 1782 else 1783 insn_buf[1] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_AX, 0, insn->off - 1); 1784 insn_buf[2] = BPF_ALU64_IMM(BPF_SUB, BPF_REG_AX, 1); 1785 insn_buf[3] = BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_AX, stack_off); 1786 cnt = 4; 1787 1788 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1789 if (!new_prog) 1790 return -ENOMEM; 1791 1792 delta += cnt - 1; 1793 env->prog = prog = new_prog; 1794 insn = new_prog->insnsi + i + delta; 1795 goto next_insn; 1796 } 1797 1798 if (insn->code != (BPF_JMP | BPF_CALL)) 1799 goto next_insn; 1800 if (insn->src_reg == BPF_PSEUDO_CALL) 1801 goto next_insn; 1802 if (insn->src_reg == BPF_PSEUDO_KFUNC_CALL) { 1803 ret = bpf_fixup_kfunc_call(env, insn, insn_buf, i + delta, &cnt); 1804 if (ret) 1805 return ret; 1806 if (cnt == 0) 1807 goto next_insn; 1808 1809 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1810 if (!new_prog) 1811 return -ENOMEM; 1812 1813 delta += cnt - 1; 1814 env->prog = prog = new_prog; 1815 insn = new_prog->insnsi + i + delta; 1816 goto next_insn; 1817 } 1818 1819 /* Skip inlining the helper call if the JIT does it. */ 1820 if (bpf_jit_inlines_helper_call(insn->imm)) 1821 goto next_insn; 1822 1823 if (insn->imm == BPF_FUNC_get_route_realm) 1824 prog->dst_needed = 1; 1825 if (insn->imm == BPF_FUNC_get_prandom_u32) 1826 bpf_user_rnd_init_once(); 1827 if (insn->imm == BPF_FUNC_override_return) 1828 prog->kprobe_override = 1; 1829 if (insn->imm == BPF_FUNC_tail_call) { 1830 /* If we tail call into other programs, we 1831 * cannot make any assumptions since they can 1832 * be replaced dynamically during runtime in 1833 * the program array. 1834 */ 1835 prog->cb_access = 1; 1836 if (!bpf_allow_tail_call_in_subprogs(env)) 1837 prog->aux->stack_depth = MAX_BPF_STACK; 1838 prog->aux->max_pkt_offset = MAX_PACKET_OFF; 1839 1840 /* mark bpf_tail_call as different opcode to avoid 1841 * conditional branch in the interpreter for every normal 1842 * call and to prevent accidental JITing by JIT compiler 1843 * that doesn't support bpf_tail_call yet 1844 */ 1845 insn->imm = 0; 1846 insn->code = BPF_JMP | BPF_TAIL_CALL; 1847 1848 aux = &env->insn_aux_data[i + delta]; 1849 if (env->bpf_capable && !prog->blinding_requested && 1850 prog->jit_requested && 1851 !bpf_map_key_poisoned(aux) && 1852 !bpf_map_ptr_poisoned(aux) && 1853 !bpf_map_ptr_unpriv(aux)) { 1854 struct bpf_jit_poke_descriptor desc = { 1855 .reason = BPF_POKE_REASON_TAIL_CALL, 1856 .tail_call.map = aux->map_ptr_state.map_ptr, 1857 .tail_call.key = bpf_map_key_immediate(aux), 1858 .insn_idx = i + delta, 1859 }; 1860 1861 ret = bpf_jit_add_poke_descriptor(prog, &desc); 1862 if (ret < 0) { 1863 verbose(env, "adding tail call poke descriptor failed\n"); 1864 return ret; 1865 } 1866 1867 insn->imm = ret + 1; 1868 goto next_insn; 1869 } 1870 1871 if (!bpf_map_ptr_unpriv(aux)) 1872 goto next_insn; 1873 1874 /* instead of changing every JIT dealing with tail_call 1875 * emit two extra insns: 1876 * if (index >= max_entries) goto out; 1877 * index &= array->index_mask; 1878 * to avoid out-of-bounds cpu speculation 1879 */ 1880 if (bpf_map_ptr_poisoned(aux)) { 1881 verbose(env, "tail_call abusing map_ptr\n"); 1882 return -EINVAL; 1883 } 1884 1885 map_ptr = aux->map_ptr_state.map_ptr; 1886 insn_buf[0] = BPF_JMP_IMM(BPF_JGE, BPF_REG_3, 1887 map_ptr->max_entries, 2); 1888 insn_buf[1] = BPF_ALU32_IMM(BPF_AND, BPF_REG_3, 1889 container_of(map_ptr, 1890 struct bpf_array, 1891 map)->index_mask); 1892 insn_buf[2] = *insn; 1893 cnt = 3; 1894 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1895 if (!new_prog) 1896 return -ENOMEM; 1897 1898 delta += cnt - 1; 1899 env->prog = prog = new_prog; 1900 insn = new_prog->insnsi + i + delta; 1901 goto next_insn; 1902 } 1903 1904 if (insn->imm == BPF_FUNC_timer_set_callback) { 1905 /* The verifier will process callback_fn as many times as necessary 1906 * with different maps and the register states prepared by 1907 * set_timer_callback_state will be accurate. 1908 * 1909 * The following use case is valid: 1910 * map1 is shared by prog1, prog2, prog3. 1911 * prog1 calls bpf_timer_init for some map1 elements 1912 * prog2 calls bpf_timer_set_callback for some map1 elements. 1913 * Those that were not bpf_timer_init-ed will return -EINVAL. 1914 * prog3 calls bpf_timer_start for some map1 elements. 1915 * Those that were not both bpf_timer_init-ed and 1916 * bpf_timer_set_callback-ed will return -EINVAL. 1917 */ 1918 struct bpf_insn ld_addrs[2] = { 1919 BPF_LD_IMM64(BPF_REG_3, (long)prog->aux), 1920 }; 1921 1922 insn_buf[0] = ld_addrs[0]; 1923 insn_buf[1] = ld_addrs[1]; 1924 insn_buf[2] = *insn; 1925 cnt = 3; 1926 1927 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1928 if (!new_prog) 1929 return -ENOMEM; 1930 1931 delta += cnt - 1; 1932 env->prog = prog = new_prog; 1933 insn = new_prog->insnsi + i + delta; 1934 goto patch_call_imm; 1935 } 1936 1937 /* bpf_per_cpu_ptr() and bpf_this_cpu_ptr() */ 1938 if (env->insn_aux_data[i + delta].call_with_percpu_alloc_ptr) { 1939 /* patch with 'r1 = *(u64 *)(r1 + 0)' since for percpu data, 1940 * bpf_mem_alloc() returns a ptr to the percpu data ptr. 1941 */ 1942 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_1, BPF_REG_1, 0); 1943 insn_buf[1] = *insn; 1944 cnt = 2; 1945 1946 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 1947 if (!new_prog) 1948 return -ENOMEM; 1949 1950 delta += cnt - 1; 1951 env->prog = prog = new_prog; 1952 insn = new_prog->insnsi + i + delta; 1953 goto patch_call_imm; 1954 } 1955 1956 /* BPF_EMIT_CALL() assumptions in some of the map_gen_lookup 1957 * and other inlining handlers are currently limited to 64 bit 1958 * only. 1959 */ 1960 if (prog->jit_requested && BITS_PER_LONG == 64 && 1961 (insn->imm == BPF_FUNC_map_lookup_elem || 1962 insn->imm == BPF_FUNC_map_update_elem || 1963 insn->imm == BPF_FUNC_map_delete_elem || 1964 insn->imm == BPF_FUNC_map_push_elem || 1965 insn->imm == BPF_FUNC_map_pop_elem || 1966 insn->imm == BPF_FUNC_map_peek_elem || 1967 insn->imm == BPF_FUNC_redirect_map || 1968 insn->imm == BPF_FUNC_for_each_map_elem || 1969 insn->imm == BPF_FUNC_map_lookup_percpu_elem)) { 1970 aux = &env->insn_aux_data[i + delta]; 1971 if (bpf_map_ptr_poisoned(aux)) 1972 goto patch_call_imm; 1973 1974 map_ptr = aux->map_ptr_state.map_ptr; 1975 ops = map_ptr->ops; 1976 if (insn->imm == BPF_FUNC_map_lookup_elem && 1977 ops->map_gen_lookup) { 1978 cnt = ops->map_gen_lookup(map_ptr, insn_buf); 1979 if (cnt == -EOPNOTSUPP) 1980 goto patch_map_ops_generic; 1981 if (cnt <= 0 || cnt >= INSN_BUF_SIZE) { 1982 verifier_bug(env, "%d insns generated for map lookup", cnt); 1983 return -EFAULT; 1984 } 1985 1986 new_prog = bpf_patch_insn_data(env, i + delta, 1987 insn_buf, cnt); 1988 if (!new_prog) 1989 return -ENOMEM; 1990 1991 delta += cnt - 1; 1992 env->prog = prog = new_prog; 1993 insn = new_prog->insnsi + i + delta; 1994 goto next_insn; 1995 } 1996 1997 BUILD_BUG_ON(!__same_type(ops->map_lookup_elem, 1998 (void *(*)(struct bpf_map *map, void *key))NULL)); 1999 BUILD_BUG_ON(!__same_type(ops->map_delete_elem, 2000 (long (*)(struct bpf_map *map, void *key))NULL)); 2001 BUILD_BUG_ON(!__same_type(ops->map_update_elem, 2002 (long (*)(struct bpf_map *map, void *key, void *value, 2003 u64 flags))NULL)); 2004 BUILD_BUG_ON(!__same_type(ops->map_push_elem, 2005 (long (*)(struct bpf_map *map, void *value, 2006 u64 flags))NULL)); 2007 BUILD_BUG_ON(!__same_type(ops->map_pop_elem, 2008 (long (*)(struct bpf_map *map, void *value))NULL)); 2009 BUILD_BUG_ON(!__same_type(ops->map_peek_elem, 2010 (long (*)(struct bpf_map *map, void *value))NULL)); 2011 BUILD_BUG_ON(!__same_type(ops->map_redirect, 2012 (long (*)(struct bpf_map *map, u64 index, u64 flags))NULL)); 2013 BUILD_BUG_ON(!__same_type(ops->map_for_each_callback, 2014 (long (*)(struct bpf_map *map, 2015 bpf_callback_t callback_fn, 2016 void *callback_ctx, 2017 u64 flags))NULL)); 2018 BUILD_BUG_ON(!__same_type(ops->map_lookup_percpu_elem, 2019 (void *(*)(struct bpf_map *map, void *key, u32 cpu))NULL)); 2020 2021 patch_map_ops_generic: 2022 switch (insn->imm) { 2023 case BPF_FUNC_map_lookup_elem: 2024 insn->imm = BPF_CALL_IMM(ops->map_lookup_elem); 2025 goto next_insn; 2026 case BPF_FUNC_map_update_elem: 2027 insn->imm = BPF_CALL_IMM(ops->map_update_elem); 2028 goto next_insn; 2029 case BPF_FUNC_map_delete_elem: 2030 insn->imm = BPF_CALL_IMM(ops->map_delete_elem); 2031 goto next_insn; 2032 case BPF_FUNC_map_push_elem: 2033 insn->imm = BPF_CALL_IMM(ops->map_push_elem); 2034 goto next_insn; 2035 case BPF_FUNC_map_pop_elem: 2036 insn->imm = BPF_CALL_IMM(ops->map_pop_elem); 2037 goto next_insn; 2038 case BPF_FUNC_map_peek_elem: 2039 insn->imm = BPF_CALL_IMM(ops->map_peek_elem); 2040 goto next_insn; 2041 case BPF_FUNC_redirect_map: 2042 insn->imm = BPF_CALL_IMM(ops->map_redirect); 2043 goto next_insn; 2044 case BPF_FUNC_for_each_map_elem: 2045 insn->imm = BPF_CALL_IMM(ops->map_for_each_callback); 2046 goto next_insn; 2047 case BPF_FUNC_map_lookup_percpu_elem: 2048 insn->imm = BPF_CALL_IMM(ops->map_lookup_percpu_elem); 2049 goto next_insn; 2050 } 2051 2052 goto patch_call_imm; 2053 } 2054 2055 /* Implement bpf_jiffies64 inline. */ 2056 if (prog->jit_requested && BITS_PER_LONG == 64 && 2057 insn->imm == BPF_FUNC_jiffies64) { 2058 struct bpf_insn ld_jiffies_addr[2] = { 2059 BPF_LD_IMM64(BPF_REG_0, 2060 (unsigned long)&jiffies), 2061 }; 2062 2063 insn_buf[0] = ld_jiffies_addr[0]; 2064 insn_buf[1] = ld_jiffies_addr[1]; 2065 insn_buf[2] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, 2066 BPF_REG_0, 0); 2067 cnt = 3; 2068 2069 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, 2070 cnt); 2071 if (!new_prog) 2072 return -ENOMEM; 2073 2074 delta += cnt - 1; 2075 env->prog = prog = new_prog; 2076 insn = new_prog->insnsi + i + delta; 2077 goto next_insn; 2078 } 2079 2080 #if defined(CONFIG_X86_64) && !defined(CONFIG_UML) 2081 /* Implement bpf_get_smp_processor_id() inline. */ 2082 if (insn->imm == BPF_FUNC_get_smp_processor_id && 2083 bpf_verifier_inlines_helper_call(env, insn->imm)) { 2084 /* BPF_FUNC_get_smp_processor_id inlining is an 2085 * optimization, so if cpu_number is ever 2086 * changed in some incompatible and hard to support 2087 * way, it's fine to back out this inlining logic 2088 */ 2089 #ifdef CONFIG_SMP 2090 insn_buf[0] = BPF_MOV64_IMM(BPF_REG_0, (u32)(unsigned long)&cpu_number); 2091 insn_buf[1] = BPF_MOV64_PERCPU_REG(BPF_REG_0, BPF_REG_0); 2092 insn_buf[2] = BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_0, 0); 2093 cnt = 3; 2094 #else 2095 insn_buf[0] = BPF_ALU32_REG(BPF_XOR, BPF_REG_0, BPF_REG_0); 2096 cnt = 1; 2097 #endif 2098 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2099 if (!new_prog) 2100 return -ENOMEM; 2101 2102 delta += cnt - 1; 2103 env->prog = prog = new_prog; 2104 insn = new_prog->insnsi + i + delta; 2105 goto next_insn; 2106 } 2107 2108 /* Implement bpf_get_current_task() and bpf_get_current_task_btf() inline. */ 2109 if ((insn->imm == BPF_FUNC_get_current_task || insn->imm == BPF_FUNC_get_current_task_btf) && 2110 bpf_verifier_inlines_helper_call(env, insn->imm)) { 2111 insn_buf[0] = BPF_MOV64_IMM(BPF_REG_0, (u32)(unsigned long)¤t_task); 2112 insn_buf[1] = BPF_MOV64_PERCPU_REG(BPF_REG_0, BPF_REG_0); 2113 insn_buf[2] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_0, 0); 2114 cnt = 3; 2115 2116 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2117 if (!new_prog) 2118 return -ENOMEM; 2119 2120 delta += cnt - 1; 2121 env->prog = prog = new_prog; 2122 insn = new_prog->insnsi + i + delta; 2123 goto next_insn; 2124 } 2125 #endif 2126 /* Implement bpf_get_func_arg inline. */ 2127 if (prog_type == BPF_PROG_TYPE_TRACING && 2128 insn->imm == BPF_FUNC_get_func_arg) { 2129 if (eatype == BPF_TRACE_RAW_TP) { 2130 int nr_args = btf_type_vlen(prog->aux->attach_func_proto); 2131 2132 /* skip 'void *__data' in btf_trace_##name() and save to reg0 */ 2133 insn_buf[0] = BPF_MOV64_IMM(BPF_REG_0, nr_args - 1); 2134 cnt = 1; 2135 } else { 2136 /* Load nr_args from ctx - 8 */ 2137 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_1, -8); 2138 insn_buf[1] = BPF_ALU64_IMM(BPF_AND, BPF_REG_0, 0xFF); 2139 cnt = 2; 2140 } 2141 insn_buf[cnt++] = BPF_JMP32_REG(BPF_JGE, BPF_REG_2, BPF_REG_0, 6); 2142 insn_buf[cnt++] = BPF_ALU64_IMM(BPF_LSH, BPF_REG_2, 3); 2143 insn_buf[cnt++] = BPF_ALU64_REG(BPF_ADD, BPF_REG_2, BPF_REG_1); 2144 insn_buf[cnt++] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_2, 0); 2145 insn_buf[cnt++] = BPF_STX_MEM(BPF_DW, BPF_REG_3, BPF_REG_0, 0); 2146 insn_buf[cnt++] = BPF_MOV64_IMM(BPF_REG_0, 0); 2147 insn_buf[cnt++] = BPF_JMP_A(1); 2148 insn_buf[cnt++] = BPF_MOV64_IMM(BPF_REG_0, -EINVAL); 2149 2150 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2151 if (!new_prog) 2152 return -ENOMEM; 2153 2154 delta += cnt - 1; 2155 env->prog = prog = new_prog; 2156 insn = new_prog->insnsi + i + delta; 2157 goto next_insn; 2158 } 2159 2160 /* Implement bpf_get_func_ret inline. */ 2161 if (prog_type == BPF_PROG_TYPE_TRACING && 2162 insn->imm == BPF_FUNC_get_func_ret) { 2163 if (eatype == BPF_TRACE_FEXIT || 2164 eatype == BPF_TRACE_FSESSION || 2165 eatype == BPF_MODIFY_RETURN) { 2166 /* Load nr_args from ctx - 8 */ 2167 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_1, -8); 2168 insn_buf[1] = BPF_ALU64_IMM(BPF_AND, BPF_REG_0, 0xFF); 2169 insn_buf[2] = BPF_ALU64_IMM(BPF_LSH, BPF_REG_0, 3); 2170 insn_buf[3] = BPF_ALU64_REG(BPF_ADD, BPF_REG_0, BPF_REG_1); 2171 insn_buf[4] = BPF_LDX_MEM(BPF_DW, BPF_REG_3, BPF_REG_0, 0); 2172 insn_buf[5] = BPF_STX_MEM(BPF_DW, BPF_REG_2, BPF_REG_3, 0); 2173 insn_buf[6] = BPF_MOV64_IMM(BPF_REG_0, 0); 2174 cnt = 7; 2175 } else { 2176 insn_buf[0] = BPF_MOV64_IMM(BPF_REG_0, -EOPNOTSUPP); 2177 cnt = 1; 2178 } 2179 2180 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2181 if (!new_prog) 2182 return -ENOMEM; 2183 2184 delta += cnt - 1; 2185 env->prog = prog = new_prog; 2186 insn = new_prog->insnsi + i + delta; 2187 goto next_insn; 2188 } 2189 2190 /* Implement get_func_arg_cnt inline. */ 2191 if (prog_type == BPF_PROG_TYPE_TRACING && 2192 insn->imm == BPF_FUNC_get_func_arg_cnt) { 2193 if (eatype == BPF_TRACE_RAW_TP) { 2194 int nr_args = btf_type_vlen(prog->aux->attach_func_proto); 2195 2196 /* skip 'void *__data' in btf_trace_##name() and save to reg0 */ 2197 insn_buf[0] = BPF_MOV64_IMM(BPF_REG_0, nr_args - 1); 2198 cnt = 1; 2199 } else { 2200 /* Load nr_args from ctx - 8 */ 2201 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_1, -8); 2202 insn_buf[1] = BPF_ALU64_IMM(BPF_AND, BPF_REG_0, 0xFF); 2203 cnt = 2; 2204 } 2205 2206 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2207 if (!new_prog) 2208 return -ENOMEM; 2209 2210 delta += cnt - 1; 2211 env->prog = prog = new_prog; 2212 insn = new_prog->insnsi + i + delta; 2213 goto next_insn; 2214 } 2215 2216 /* Implement bpf_get_func_ip inline. */ 2217 if (prog_type == BPF_PROG_TYPE_TRACING && 2218 insn->imm == BPF_FUNC_get_func_ip) { 2219 /* Load IP address from ctx - 16 */ 2220 insn_buf[0] = BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_1, -16); 2221 2222 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, 1); 2223 if (!new_prog) 2224 return -ENOMEM; 2225 2226 env->prog = prog = new_prog; 2227 insn = new_prog->insnsi + i + delta; 2228 goto next_insn; 2229 } 2230 2231 /* Implement bpf_get_branch_snapshot inline. */ 2232 if (IS_ENABLED(CONFIG_PERF_EVENTS) && 2233 prog->jit_requested && BITS_PER_LONG == 64 && 2234 insn->imm == BPF_FUNC_get_branch_snapshot) { 2235 /* We are dealing with the following func protos: 2236 * u64 bpf_get_branch_snapshot(void *buf, u32 size, u64 flags); 2237 * int perf_snapshot_branch_stack(struct perf_branch_entry *entries, u32 cnt); 2238 */ 2239 const u32 br_entry_size = sizeof(struct perf_branch_entry); 2240 2241 /* struct perf_branch_entry is part of UAPI and is 2242 * used as an array element, so extremely unlikely to 2243 * ever grow or shrink 2244 */ 2245 BUILD_BUG_ON(br_entry_size != 24); 2246 2247 /* if (unlikely(flags)) return -EINVAL */ 2248 insn_buf[0] = BPF_JMP_IMM(BPF_JNE, BPF_REG_3, 0, 7); 2249 2250 /* Transform size (bytes) into number of entries (cnt = size / 24). 2251 * But to avoid expensive division instruction, we implement 2252 * divide-by-3 through multiplication, followed by further 2253 * division by 8 through 3-bit right shift. 2254 * Refer to book "Hacker's Delight, 2nd ed." by Henry S. Warren, Jr., 2255 * p. 227, chapter "Unsigned Division by 3" for details and proofs. 2256 * 2257 * N / 3 <=> M * N / 2^33, where M = (2^33 + 1) / 3 = 0xaaaaaaab. 2258 */ 2259 insn_buf[1] = BPF_MOV32_IMM(BPF_REG_0, 0xaaaaaaab); 2260 insn_buf[2] = BPF_ALU64_REG(BPF_MUL, BPF_REG_2, BPF_REG_0); 2261 insn_buf[3] = BPF_ALU64_IMM(BPF_RSH, BPF_REG_2, 36); 2262 2263 /* call perf_snapshot_branch_stack implementation */ 2264 insn_buf[4] = BPF_EMIT_CALL(static_call_query(perf_snapshot_branch_stack)); 2265 /* if (entry_cnt == 0) return -ENOENT */ 2266 insn_buf[5] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 0, 4); 2267 /* return entry_cnt * sizeof(struct perf_branch_entry) */ 2268 insn_buf[6] = BPF_ALU32_IMM(BPF_MUL, BPF_REG_0, br_entry_size); 2269 insn_buf[7] = BPF_JMP_A(3); 2270 /* return -EINVAL; */ 2271 insn_buf[8] = BPF_MOV64_IMM(BPF_REG_0, -EINVAL); 2272 insn_buf[9] = BPF_JMP_A(1); 2273 /* return -ENOENT; */ 2274 insn_buf[10] = BPF_MOV64_IMM(BPF_REG_0, -ENOENT); 2275 cnt = 11; 2276 2277 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2278 if (!new_prog) 2279 return -ENOMEM; 2280 2281 delta += cnt - 1; 2282 env->prog = prog = new_prog; 2283 insn = new_prog->insnsi + i + delta; 2284 goto next_insn; 2285 } 2286 2287 /* Implement bpf_kptr_xchg inline */ 2288 if (prog->jit_requested && BITS_PER_LONG == 64 && 2289 insn->imm == BPF_FUNC_kptr_xchg && 2290 bpf_jit_supports_ptr_xchg()) { 2291 insn_buf[0] = BPF_MOV64_REG(BPF_REG_0, BPF_REG_2); 2292 insn_buf[1] = BPF_ATOMIC_OP(BPF_DW, BPF_XCHG, BPF_REG_1, BPF_REG_0, 0); 2293 cnt = 2; 2294 2295 new_prog = bpf_patch_insn_data(env, i + delta, insn_buf, cnt); 2296 if (!new_prog) 2297 return -ENOMEM; 2298 2299 delta += cnt - 1; 2300 env->prog = prog = new_prog; 2301 insn = new_prog->insnsi + i + delta; 2302 goto next_insn; 2303 } 2304 patch_call_imm: 2305 fn = env->ops->get_func_proto(insn->imm, env->prog); 2306 /* all functions that have prototype and verifier allowed 2307 * programs to call them, must be real in-kernel functions 2308 */ 2309 if (!fn->func) { 2310 verifier_bug(env, 2311 "not inlined functions %s#%d is missing func", 2312 func_id_name(insn->imm), insn->imm); 2313 return -EFAULT; 2314 } 2315 insn->imm = fn->func - __bpf_call_base; 2316 next_insn: 2317 if (subprogs[cur_subprog + 1].start == i + delta + 1) { 2318 subprogs[cur_subprog].stack_depth += stack_depth_extra; 2319 subprogs[cur_subprog].stack_extra = stack_depth_extra; 2320 2321 stack_depth = subprogs[cur_subprog].stack_depth; 2322 if (stack_depth > MAX_BPF_STACK && !prog->jit_requested) { 2323 verbose(env, "stack size %d(extra %d) is too large\n", 2324 stack_depth, stack_depth_extra); 2325 return -EINVAL; 2326 } 2327 cur_subprog++; 2328 stack_depth = subprogs[cur_subprog].stack_depth; 2329 stack_depth_extra = 0; 2330 } 2331 i++; 2332 insn++; 2333 } 2334 2335 env->prog->aux->stack_depth = subprogs[0].stack_depth; 2336 for (i = 0; i < env->subprog_cnt; i++) { 2337 int delta = bpf_jit_supports_timed_may_goto() ? 2 : 1; 2338 int subprog_start = subprogs[i].start; 2339 int stack_slots = subprogs[i].stack_extra / 8; 2340 int slots = delta, cnt = 0; 2341 2342 if (!stack_slots) 2343 continue; 2344 /* We need two slots in case timed may_goto is supported. */ 2345 if (stack_slots > slots) { 2346 verifier_bug(env, "stack_slots supports may_goto only"); 2347 return -EFAULT; 2348 } 2349 2350 stack_depth = subprogs[i].stack_depth; 2351 if (bpf_jit_supports_timed_may_goto()) { 2352 insn_buf[cnt++] = BPF_ST_MEM(BPF_DW, BPF_REG_FP, -stack_depth, 2353 BPF_MAX_TIMED_LOOPS); 2354 insn_buf[cnt++] = BPF_ST_MEM(BPF_DW, BPF_REG_FP, -stack_depth + 8, 0); 2355 } else { 2356 /* Add ST insn to subprog prologue to init extra stack */ 2357 insn_buf[cnt++] = BPF_ST_MEM(BPF_DW, BPF_REG_FP, -stack_depth, 2358 BPF_MAX_LOOPS); 2359 } 2360 /* Copy first actual insn to preserve it */ 2361 insn_buf[cnt++] = env->prog->insnsi[subprog_start]; 2362 2363 new_prog = bpf_patch_insn_data(env, subprog_start, insn_buf, cnt); 2364 if (!new_prog) 2365 return -ENOMEM; 2366 env->prog = prog = new_prog; 2367 /* 2368 * If may_goto is a first insn of a prog there could be a jmp 2369 * insn that points to it, hence adjust all such jmps to point 2370 * to insn after BPF_ST that inits may_goto count. 2371 * Adjustment will succeed because bpf_patch_insn_data() didn't fail. 2372 */ 2373 WARN_ON(adjust_jmp_off(env->prog, subprog_start, delta)); 2374 } 2375 2376 /* Since poke tab is now finalized, publish aux to tracker. */ 2377 for (i = 0; i < prog->aux->size_poke_tab; i++) { 2378 map_ptr = prog->aux->poke_tab[i].tail_call.map; 2379 if (!map_ptr->ops->map_poke_track || 2380 !map_ptr->ops->map_poke_untrack || 2381 !map_ptr->ops->map_poke_run) { 2382 verifier_bug(env, "poke tab is misconfigured"); 2383 return -EFAULT; 2384 } 2385 2386 ret = map_ptr->ops->map_poke_track(map_ptr, prog->aux); 2387 if (ret < 0) { 2388 verbose(env, "tracking tail call prog failed\n"); 2389 return ret; 2390 } 2391 } 2392 2393 ret = sort_kfunc_descs_by_imm_off(env); 2394 if (ret) 2395 return ret; 2396 2397 return 0; 2398 } 2399 2400 static struct bpf_prog *inline_bpf_loop(struct bpf_verifier_env *env, 2401 int position, 2402 s32 stack_base, 2403 u32 callback_subprogno, 2404 u32 *total_cnt) 2405 { 2406 s32 r6_offset = stack_base + 0 * BPF_REG_SIZE; 2407 s32 r7_offset = stack_base + 1 * BPF_REG_SIZE; 2408 s32 r8_offset = stack_base + 2 * BPF_REG_SIZE; 2409 int reg_loop_max = BPF_REG_6; 2410 int reg_loop_cnt = BPF_REG_7; 2411 int reg_loop_ctx = BPF_REG_8; 2412 2413 struct bpf_insn *insn_buf = env->insn_buf; 2414 struct bpf_prog *new_prog; 2415 u32 callback_start; 2416 u32 call_insn_offset; 2417 s32 callback_offset; 2418 u32 cnt = 0; 2419 2420 /* This represents an inlined version of bpf_iter.c:bpf_loop, 2421 * be careful to modify this code in sync. 2422 */ 2423 2424 /* Return error and jump to the end of the patch if 2425 * expected number of iterations is too big. 2426 */ 2427 insn_buf[cnt++] = BPF_JMP_IMM(BPF_JLE, BPF_REG_1, BPF_MAX_LOOPS, 2); 2428 insn_buf[cnt++] = BPF_MOV32_IMM(BPF_REG_0, -E2BIG); 2429 insn_buf[cnt++] = BPF_JMP_IMM(BPF_JA, 0, 0, 16); 2430 /* spill R6, R7, R8 to use these as loop vars */ 2431 insn_buf[cnt++] = BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_6, r6_offset); 2432 insn_buf[cnt++] = BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_7, r7_offset); 2433 insn_buf[cnt++] = BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_8, r8_offset); 2434 /* initialize loop vars */ 2435 insn_buf[cnt++] = BPF_MOV64_REG(reg_loop_max, BPF_REG_1); 2436 insn_buf[cnt++] = BPF_MOV32_IMM(reg_loop_cnt, 0); 2437 insn_buf[cnt++] = BPF_MOV64_REG(reg_loop_ctx, BPF_REG_3); 2438 /* loop header, 2439 * if reg_loop_cnt >= reg_loop_max skip the loop body 2440 */ 2441 insn_buf[cnt++] = BPF_JMP_REG(BPF_JGE, reg_loop_cnt, reg_loop_max, 5); 2442 /* callback call, 2443 * correct callback offset would be set after patching 2444 */ 2445 insn_buf[cnt++] = BPF_MOV64_REG(BPF_REG_1, reg_loop_cnt); 2446 insn_buf[cnt++] = BPF_MOV64_REG(BPF_REG_2, reg_loop_ctx); 2447 insn_buf[cnt++] = BPF_CALL_REL(0); 2448 /* increment loop counter */ 2449 insn_buf[cnt++] = BPF_ALU64_IMM(BPF_ADD, reg_loop_cnt, 1); 2450 /* jump to loop header if callback returned 0 */ 2451 insn_buf[cnt++] = BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 0, -6); 2452 /* return value of bpf_loop, 2453 * set R0 to the number of iterations 2454 */ 2455 insn_buf[cnt++] = BPF_MOV64_REG(BPF_REG_0, reg_loop_cnt); 2456 /* restore original values of R6, R7, R8 */ 2457 insn_buf[cnt++] = BPF_LDX_MEM(BPF_DW, BPF_REG_6, BPF_REG_10, r6_offset); 2458 insn_buf[cnt++] = BPF_LDX_MEM(BPF_DW, BPF_REG_7, BPF_REG_10, r7_offset); 2459 insn_buf[cnt++] = BPF_LDX_MEM(BPF_DW, BPF_REG_8, BPF_REG_10, r8_offset); 2460 2461 *total_cnt = cnt; 2462 new_prog = bpf_patch_insn_data(env, position, insn_buf, cnt); 2463 if (!new_prog) 2464 return new_prog; 2465 2466 /* callback start is known only after patching */ 2467 callback_start = env->subprog_info[callback_subprogno].start; 2468 /* Note: insn_buf[12] is an offset of BPF_CALL_REL instruction */ 2469 call_insn_offset = position + 12; 2470 callback_offset = callback_start - call_insn_offset - 1; 2471 new_prog->insnsi[call_insn_offset].imm = callback_offset; 2472 2473 return new_prog; 2474 } 2475 2476 static bool is_bpf_loop_call(struct bpf_insn *insn) 2477 { 2478 return insn->code == (BPF_JMP | BPF_CALL) && 2479 insn->src_reg == 0 && 2480 insn->imm == BPF_FUNC_loop; 2481 } 2482 2483 /* For all sub-programs in the program (including main) check 2484 * insn_aux_data to see if there are bpf_loop calls that require 2485 * inlining. If such calls are found the calls are replaced with a 2486 * sequence of instructions produced by `inline_bpf_loop` function and 2487 * subprog stack_depth is increased by the size of 3 registers. 2488 * This stack space is used to spill values of the R6, R7, R8. These 2489 * registers are used to store the loop bound, counter and context 2490 * variables. 2491 */ 2492 int bpf_optimize_bpf_loop(struct bpf_verifier_env *env) 2493 { 2494 struct bpf_subprog_info *subprogs = env->subprog_info; 2495 int i, cur_subprog = 0, cnt, delta = 0; 2496 struct bpf_insn *insn = env->prog->insnsi; 2497 int insn_cnt = env->prog->len; 2498 u16 stack_depth = subprogs[cur_subprog].stack_depth; 2499 u16 stack_depth_roundup = round_up(stack_depth, 8) - stack_depth; 2500 u16 stack_depth_extra = 0; 2501 2502 for (i = 0; i < insn_cnt; i++, insn++) { 2503 struct bpf_loop_inline_state *inline_state = 2504 &env->insn_aux_data[i + delta].loop_inline_state; 2505 2506 if (is_bpf_loop_call(insn) && inline_state->fit_for_inline) { 2507 struct bpf_prog *new_prog; 2508 2509 stack_depth_extra = BPF_REG_SIZE * 3 + stack_depth_roundup; 2510 new_prog = inline_bpf_loop(env, 2511 i + delta, 2512 -(stack_depth + stack_depth_extra), 2513 inline_state->callback_subprogno, 2514 &cnt); 2515 if (!new_prog) 2516 return -ENOMEM; 2517 2518 delta += cnt - 1; 2519 env->prog = new_prog; 2520 insn = new_prog->insnsi + i + delta; 2521 } 2522 2523 if (subprogs[cur_subprog + 1].start == i + delta + 1) { 2524 subprogs[cur_subprog].stack_depth += stack_depth_extra; 2525 cur_subprog++; 2526 stack_depth = subprogs[cur_subprog].stack_depth; 2527 stack_depth_roundup = round_up(stack_depth, 8) - stack_depth; 2528 stack_depth_extra = 0; 2529 } 2530 } 2531 2532 env->prog->aux->stack_depth = env->subprog_info[0].stack_depth; 2533 2534 return 0; 2535 } 2536 2537 /* Remove unnecessary spill/fill pairs, members of fastcall pattern, 2538 * adjust subprograms stack depth when possible. 2539 */ 2540 int bpf_remove_fastcall_spills_fills(struct bpf_verifier_env *env) 2541 { 2542 struct bpf_subprog_info *subprog = env->subprog_info; 2543 struct bpf_insn_aux_data *aux = env->insn_aux_data; 2544 struct bpf_insn *insn = env->prog->insnsi; 2545 int insn_cnt = env->prog->len; 2546 u32 spills_num; 2547 bool modified = false; 2548 int i, j; 2549 2550 for (i = 0; i < insn_cnt; i++, insn++) { 2551 if (aux[i].fastcall_spills_num > 0) { 2552 spills_num = aux[i].fastcall_spills_num; 2553 /* NOPs would be removed by opt_remove_nops() */ 2554 for (j = 1; j <= spills_num; ++j) { 2555 *(insn - j) = NOP; 2556 *(insn + j) = NOP; 2557 } 2558 modified = true; 2559 } 2560 if ((subprog + 1)->start == i + 1) { 2561 if (modified && !subprog->keep_fastcall_stack) 2562 subprog->stack_depth = -subprog->fastcall_stack_off; 2563 subprog++; 2564 modified = false; 2565 } 2566 } 2567 2568 return 0; 2569 } 2570 2571