Commit 61bd5218 authored by Jakub Kicinski's avatar Jakub Kicinski Committed by David S. Miller

bpf: move global verifier log into verifier environment

The biggest piece of global state protected by the verifier lock
is the verifier_log.  Move that log to struct bpf_verifier_env.
struct bpf_verifier_env has to be passed now to all invocations
of verbose().
Signed-off-by: default avatarJakub Kicinski <jakub.kicinski@netronome.com>
Reviewed-by: default avatarSimon Horman <simon.horman@netronome.com>
Acked-by: default avatarAlexei Starovoitov <ast@kernel.org>
Acked-by: default avatarDaniel Borkmann <daniel@iogearbox.net>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent e7bf8249
...@@ -152,6 +152,8 @@ struct bpf_verifier_env { ...@@ -152,6 +152,8 @@ struct bpf_verifier_env {
bool allow_ptr_leaks; bool allow_ptr_leaks;
bool seen_direct_write; bool seen_direct_write;
struct bpf_insn_aux_data *insn_aux_data; /* array of per-insn state */ struct bpf_insn_aux_data *insn_aux_data; /* array of per-insn state */
struct bpf_verifer_log log;
}; };
int bpf_analyzer(struct bpf_prog *prog, const struct bpf_ext_analyzer_ops *ops, int bpf_analyzer(struct bpf_prog *prog, const struct bpf_ext_analyzer_ops *ops,
......
...@@ -153,20 +153,16 @@ struct bpf_call_arg_meta { ...@@ -153,20 +153,16 @@ struct bpf_call_arg_meta {
int access_size; int access_size;
}; };
/* verbose verifier prints what it's seeing
* bpf_check() is called under lock, so no race to access these global vars
*/
static struct bpf_verifer_log verifier_log;
static DEFINE_MUTEX(bpf_verifier_lock); static DEFINE_MUTEX(bpf_verifier_lock);
/* log_level controls verbosity level of eBPF verifier. /* log_level controls verbosity level of eBPF verifier.
* verbose() is used to dump the verification trace to the log, so the user * verbose() is used to dump the verification trace to the log, so the user
* can figure out what's wrong with the program * can figure out what's wrong with the program
*/ */
static __printf(1, 2) void verbose(const char *fmt, ...) static __printf(2, 3) void verbose(struct bpf_verifier_env *env,
const char *fmt, ...)
{ {
struct bpf_verifer_log *log = &verifier_log; struct bpf_verifer_log *log = &env->log;
va_list args; va_list args;
if (!log->level || bpf_verifier_log_full(log)) if (!log->level || bpf_verifier_log_full(log))
...@@ -214,7 +210,8 @@ static const char *func_id_name(int id) ...@@ -214,7 +210,8 @@ static const char *func_id_name(int id)
return "unknown"; return "unknown";
} }
static void print_verifier_state(struct bpf_verifier_state *state) static void print_verifier_state(struct bpf_verifier_env *env,
struct bpf_verifier_state *state)
{ {
struct bpf_reg_state *reg; struct bpf_reg_state *reg;
enum bpf_reg_type t; enum bpf_reg_type t;
...@@ -225,21 +222,21 @@ static void print_verifier_state(struct bpf_verifier_state *state) ...@@ -225,21 +222,21 @@ static void print_verifier_state(struct bpf_verifier_state *state)
t = reg->type; t = reg->type;
if (t == NOT_INIT) if (t == NOT_INIT)
continue; continue;
verbose(" R%d=%s", i, reg_type_str[t]); verbose(env, " R%d=%s", i, reg_type_str[t]);
if ((t == SCALAR_VALUE || t == PTR_TO_STACK) && if ((t == SCALAR_VALUE || t == PTR_TO_STACK) &&
tnum_is_const(reg->var_off)) { tnum_is_const(reg->var_off)) {
/* reg->off should be 0 for SCALAR_VALUE */ /* reg->off should be 0 for SCALAR_VALUE */
verbose("%lld", reg->var_off.value + reg->off); verbose(env, "%lld", reg->var_off.value + reg->off);
} else { } else {
verbose("(id=%d", reg->id); verbose(env, "(id=%d", reg->id);
if (t != SCALAR_VALUE) if (t != SCALAR_VALUE)
verbose(",off=%d", reg->off); verbose(env, ",off=%d", reg->off);
if (type_is_pkt_pointer(t)) if (type_is_pkt_pointer(t))
verbose(",r=%d", reg->range); verbose(env, ",r=%d", reg->range);
else if (t == CONST_PTR_TO_MAP || else if (t == CONST_PTR_TO_MAP ||
t == PTR_TO_MAP_VALUE || t == PTR_TO_MAP_VALUE ||
t == PTR_TO_MAP_VALUE_OR_NULL) t == PTR_TO_MAP_VALUE_OR_NULL)
verbose(",ks=%d,vs=%d", verbose(env, ",ks=%d,vs=%d",
reg->map_ptr->key_size, reg->map_ptr->key_size,
reg->map_ptr->value_size); reg->map_ptr->value_size);
if (tnum_is_const(reg->var_off)) { if (tnum_is_const(reg->var_off)) {
...@@ -247,38 +244,38 @@ static void print_verifier_state(struct bpf_verifier_state *state) ...@@ -247,38 +244,38 @@ static void print_verifier_state(struct bpf_verifier_state *state)
* could be a pointer whose offset is too big * could be a pointer whose offset is too big
* for reg->off * for reg->off
*/ */
verbose(",imm=%llx", reg->var_off.value); verbose(env, ",imm=%llx", reg->var_off.value);
} else { } else {
if (reg->smin_value != reg->umin_value && if (reg->smin_value != reg->umin_value &&
reg->smin_value != S64_MIN) reg->smin_value != S64_MIN)
verbose(",smin_value=%lld", verbose(env, ",smin_value=%lld",
(long long)reg->smin_value); (long long)reg->smin_value);
if (reg->smax_value != reg->umax_value && if (reg->smax_value != reg->umax_value &&
reg->smax_value != S64_MAX) reg->smax_value != S64_MAX)
verbose(",smax_value=%lld", verbose(env, ",smax_value=%lld",
(long long)reg->smax_value); (long long)reg->smax_value);
if (reg->umin_value != 0) if (reg->umin_value != 0)
verbose(",umin_value=%llu", verbose(env, ",umin_value=%llu",
(unsigned long long)reg->umin_value); (unsigned long long)reg->umin_value);
if (reg->umax_value != U64_MAX) if (reg->umax_value != U64_MAX)
verbose(",umax_value=%llu", verbose(env, ",umax_value=%llu",
(unsigned long long)reg->umax_value); (unsigned long long)reg->umax_value);
if (!tnum_is_unknown(reg->var_off)) { if (!tnum_is_unknown(reg->var_off)) {
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose(",var_off=%s", tn_buf); verbose(env, ",var_off=%s", tn_buf);
} }
} }
verbose(")"); verbose(env, ")");
} }
} }
for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) { for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) {
if (state->stack_slot_type[i] == STACK_SPILL) if (state->stack_slot_type[i] == STACK_SPILL)
verbose(" fp%d=%s", -MAX_BPF_STACK + i, verbose(env, " fp%d=%s", -MAX_BPF_STACK + i,
reg_type_str[state->spilled_regs[i / BPF_REG_SIZE].type]); reg_type_str[state->spilled_regs[i / BPF_REG_SIZE].type]);
} }
verbose("\n"); verbose(env, "\n");
} }
static const char *const bpf_class_string[] = { static const char *const bpf_class_string[] = {
...@@ -333,15 +330,15 @@ static const char *const bpf_jmp_string[16] = { ...@@ -333,15 +330,15 @@ static const char *const bpf_jmp_string[16] = {
[BPF_EXIT >> 4] = "exit", [BPF_EXIT >> 4] = "exit",
}; };
static void print_bpf_end_insn(const struct bpf_verifier_env *env, static void print_bpf_end_insn(struct bpf_verifier_env *env,
const struct bpf_insn *insn) const struct bpf_insn *insn)
{ {
verbose("(%02x) r%d = %s%d r%d\n", insn->code, insn->dst_reg, verbose(env, "(%02x) r%d = %s%d r%d\n", insn->code, insn->dst_reg,
BPF_SRC(insn->code) == BPF_TO_BE ? "be" : "le", BPF_SRC(insn->code) == BPF_TO_BE ? "be" : "le",
insn->imm, insn->dst_reg); insn->imm, insn->dst_reg);
} }
static void print_bpf_insn(const struct bpf_verifier_env *env, static void print_bpf_insn(struct bpf_verifier_env *env,
const struct bpf_insn *insn) const struct bpf_insn *insn)
{ {
u8 class = BPF_CLASS(insn->code); u8 class = BPF_CLASS(insn->code);
...@@ -349,23 +346,23 @@ static void print_bpf_insn(const struct bpf_verifier_env *env, ...@@ -349,23 +346,23 @@ static void print_bpf_insn(const struct bpf_verifier_env *env,
if (class == BPF_ALU || class == BPF_ALU64) { if (class == BPF_ALU || class == BPF_ALU64) {
if (BPF_OP(insn->code) == BPF_END) { if (BPF_OP(insn->code) == BPF_END) {
if (class == BPF_ALU64) if (class == BPF_ALU64)
verbose("BUG_alu64_%02x\n", insn->code); verbose(env, "BUG_alu64_%02x\n", insn->code);
else else
print_bpf_end_insn(env, insn); print_bpf_end_insn(env, insn);
} else if (BPF_OP(insn->code) == BPF_NEG) { } else if (BPF_OP(insn->code) == BPF_NEG) {
verbose("(%02x) r%d = %s-r%d\n", verbose(env, "(%02x) r%d = %s-r%d\n",
insn->code, insn->dst_reg, insn->code, insn->dst_reg,
class == BPF_ALU ? "(u32) " : "", class == BPF_ALU ? "(u32) " : "",
insn->dst_reg); insn->dst_reg);
} else if (BPF_SRC(insn->code) == BPF_X) { } else if (BPF_SRC(insn->code) == BPF_X) {
verbose("(%02x) %sr%d %s %sr%d\n", verbose(env, "(%02x) %sr%d %s %sr%d\n",
insn->code, class == BPF_ALU ? "(u32) " : "", insn->code, class == BPF_ALU ? "(u32) " : "",
insn->dst_reg, insn->dst_reg,
bpf_alu_string[BPF_OP(insn->code) >> 4], bpf_alu_string[BPF_OP(insn->code) >> 4],
class == BPF_ALU ? "(u32) " : "", class == BPF_ALU ? "(u32) " : "",
insn->src_reg); insn->src_reg);
} else { } else {
verbose("(%02x) %sr%d %s %s%d\n", verbose(env, "(%02x) %sr%d %s %s%d\n",
insn->code, class == BPF_ALU ? "(u32) " : "", insn->code, class == BPF_ALU ? "(u32) " : "",
insn->dst_reg, insn->dst_reg,
bpf_alu_string[BPF_OP(insn->code) >> 4], bpf_alu_string[BPF_OP(insn->code) >> 4],
...@@ -374,46 +371,46 @@ static void print_bpf_insn(const struct bpf_verifier_env *env, ...@@ -374,46 +371,46 @@ static void print_bpf_insn(const struct bpf_verifier_env *env,
} }
} else if (class == BPF_STX) { } else if (class == BPF_STX) {
if (BPF_MODE(insn->code) == BPF_MEM) if (BPF_MODE(insn->code) == BPF_MEM)
verbose("(%02x) *(%s *)(r%d %+d) = r%d\n", verbose(env, "(%02x) *(%s *)(r%d %+d) = r%d\n",
insn->code, insn->code,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->dst_reg, insn->dst_reg,
insn->off, insn->src_reg); insn->off, insn->src_reg);
else if (BPF_MODE(insn->code) == BPF_XADD) else if (BPF_MODE(insn->code) == BPF_XADD)
verbose("(%02x) lock *(%s *)(r%d %+d) += r%d\n", verbose(env, "(%02x) lock *(%s *)(r%d %+d) += r%d\n",
insn->code, insn->code,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->dst_reg, insn->off, insn->dst_reg, insn->off,
insn->src_reg); insn->src_reg);
else else
verbose("BUG_%02x\n", insn->code); verbose(env, "BUG_%02x\n", insn->code);
} else if (class == BPF_ST) { } else if (class == BPF_ST) {
if (BPF_MODE(insn->code) != BPF_MEM) { if (BPF_MODE(insn->code) != BPF_MEM) {
verbose("BUG_st_%02x\n", insn->code); verbose(env, "BUG_st_%02x\n", insn->code);
return; return;
} }
verbose("(%02x) *(%s *)(r%d %+d) = %d\n", verbose(env, "(%02x) *(%s *)(r%d %+d) = %d\n",
insn->code, insn->code,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->dst_reg, insn->dst_reg,
insn->off, insn->imm); insn->off, insn->imm);
} else if (class == BPF_LDX) { } else if (class == BPF_LDX) {
if (BPF_MODE(insn->code) != BPF_MEM) { if (BPF_MODE(insn->code) != BPF_MEM) {
verbose("BUG_ldx_%02x\n", insn->code); verbose(env, "BUG_ldx_%02x\n", insn->code);
return; return;
} }
verbose("(%02x) r%d = *(%s *)(r%d %+d)\n", verbose(env, "(%02x) r%d = *(%s *)(r%d %+d)\n",
insn->code, insn->dst_reg, insn->code, insn->dst_reg,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->src_reg, insn->off); insn->src_reg, insn->off);
} else if (class == BPF_LD) { } else if (class == BPF_LD) {
if (BPF_MODE(insn->code) == BPF_ABS) { if (BPF_MODE(insn->code) == BPF_ABS) {
verbose("(%02x) r0 = *(%s *)skb[%d]\n", verbose(env, "(%02x) r0 = *(%s *)skb[%d]\n",
insn->code, insn->code,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->imm); insn->imm);
} else if (BPF_MODE(insn->code) == BPF_IND) { } else if (BPF_MODE(insn->code) == BPF_IND) {
verbose("(%02x) r0 = *(%s *)skb[r%d + %d]\n", verbose(env, "(%02x) r0 = *(%s *)skb[r%d + %d]\n",
insn->code, insn->code,
bpf_ldst_string[BPF_SIZE(insn->code) >> 3], bpf_ldst_string[BPF_SIZE(insn->code) >> 3],
insn->src_reg, insn->imm); insn->src_reg, insn->imm);
...@@ -428,36 +425,37 @@ static void print_bpf_insn(const struct bpf_verifier_env *env, ...@@ -428,36 +425,37 @@ static void print_bpf_insn(const struct bpf_verifier_env *env,
if (map_ptr && !env->allow_ptr_leaks) if (map_ptr && !env->allow_ptr_leaks)
imm = 0; imm = 0;
verbose("(%02x) r%d = 0x%llx\n", insn->code, verbose(env, "(%02x) r%d = 0x%llx\n", insn->code,
insn->dst_reg, (unsigned long long)imm); insn->dst_reg, (unsigned long long)imm);
} else { } else {
verbose("BUG_ld_%02x\n", insn->code); verbose(env, "BUG_ld_%02x\n", insn->code);
return; return;
} }
} else if (class == BPF_JMP) { } else if (class == BPF_JMP) {
u8 opcode = BPF_OP(insn->code); u8 opcode = BPF_OP(insn->code);
if (opcode == BPF_CALL) { if (opcode == BPF_CALL) {
verbose("(%02x) call %s#%d\n", insn->code, verbose(env, "(%02x) call %s#%d\n", insn->code,
func_id_name(insn->imm), insn->imm); func_id_name(insn->imm), insn->imm);
} else if (insn->code == (BPF_JMP | BPF_JA)) { } else if (insn->code == (BPF_JMP | BPF_JA)) {
verbose("(%02x) goto pc%+d\n", verbose(env, "(%02x) goto pc%+d\n",
insn->code, insn->off); insn->code, insn->off);
} else if (insn->code == (BPF_JMP | BPF_EXIT)) { } else if (insn->code == (BPF_JMP | BPF_EXIT)) {
verbose("(%02x) exit\n", insn->code); verbose(env, "(%02x) exit\n", insn->code);
} else if (BPF_SRC(insn->code) == BPF_X) { } else if (BPF_SRC(insn->code) == BPF_X) {
verbose("(%02x) if r%d %s r%d goto pc%+d\n", verbose(env, "(%02x) if r%d %s r%d goto pc%+d\n",
insn->code, insn->dst_reg, insn->code, insn->dst_reg,
bpf_jmp_string[BPF_OP(insn->code) >> 4], bpf_jmp_string[BPF_OP(insn->code) >> 4],
insn->src_reg, insn->off); insn->src_reg, insn->off);
} else { } else {
verbose("(%02x) if r%d %s 0x%x goto pc%+d\n", verbose(env, "(%02x) if r%d %s 0x%x goto pc%+d\n",
insn->code, insn->dst_reg, insn->code, insn->dst_reg,
bpf_jmp_string[BPF_OP(insn->code) >> 4], bpf_jmp_string[BPF_OP(insn->code) >> 4],
insn->imm, insn->off); insn->imm, insn->off);
} }
} else { } else {
verbose("(%02x) %s\n", insn->code, bpf_class_string[class]); verbose(env, "(%02x) %s\n",
insn->code, bpf_class_string[class]);
} }
} }
...@@ -496,7 +494,7 @@ static struct bpf_verifier_state *push_stack(struct bpf_verifier_env *env, ...@@ -496,7 +494,7 @@ static struct bpf_verifier_state *push_stack(struct bpf_verifier_env *env,
env->head = elem; env->head = elem;
env->stack_size++; env->stack_size++;
if (env->stack_size > BPF_COMPLEXITY_LIMIT_STACK) { if (env->stack_size > BPF_COMPLEXITY_LIMIT_STACK) {
verbose("BPF program is too complex\n"); verbose(env, "BPF program is too complex\n");
goto err; goto err;
} }
return &elem->st; return &elem->st;
...@@ -534,10 +532,11 @@ static void __mark_reg_known_zero(struct bpf_reg_state *reg) ...@@ -534,10 +532,11 @@ static void __mark_reg_known_zero(struct bpf_reg_state *reg)
__mark_reg_known(reg, 0); __mark_reg_known(reg, 0);
} }
static void mark_reg_known_zero(struct bpf_reg_state *regs, u32 regno) static void mark_reg_known_zero(struct bpf_verifier_env *env,
struct bpf_reg_state *regs, u32 regno)
{ {
if (WARN_ON(regno >= MAX_BPF_REG)) { if (WARN_ON(regno >= MAX_BPF_REG)) {
verbose("mark_reg_known_zero(regs, %u)\n", regno); verbose(env, "mark_reg_known_zero(regs, %u)\n", regno);
/* Something bad happened, let's kill all regs */ /* Something bad happened, let's kill all regs */
for (regno = 0; regno < MAX_BPF_REG; regno++) for (regno = 0; regno < MAX_BPF_REG; regno++)
__mark_reg_not_init(regs + regno); __mark_reg_not_init(regs + regno);
...@@ -647,10 +646,11 @@ static void __mark_reg_unknown(struct bpf_reg_state *reg) ...@@ -647,10 +646,11 @@ static void __mark_reg_unknown(struct bpf_reg_state *reg)
__mark_reg_unbounded(reg); __mark_reg_unbounded(reg);
} }
static void mark_reg_unknown(struct bpf_reg_state *regs, u32 regno) static void mark_reg_unknown(struct bpf_verifier_env *env,
struct bpf_reg_state *regs, u32 regno)
{ {
if (WARN_ON(regno >= MAX_BPF_REG)) { if (WARN_ON(regno >= MAX_BPF_REG)) {
verbose("mark_reg_unknown(regs, %u)\n", regno); verbose(env, "mark_reg_unknown(regs, %u)\n", regno);
/* Something bad happened, let's kill all regs */ /* Something bad happened, let's kill all regs */
for (regno = 0; regno < MAX_BPF_REG; regno++) for (regno = 0; regno < MAX_BPF_REG; regno++)
__mark_reg_not_init(regs + regno); __mark_reg_not_init(regs + regno);
...@@ -665,10 +665,11 @@ static void __mark_reg_not_init(struct bpf_reg_state *reg) ...@@ -665,10 +665,11 @@ static void __mark_reg_not_init(struct bpf_reg_state *reg)
reg->type = NOT_INIT; reg->type = NOT_INIT;
} }
static void mark_reg_not_init(struct bpf_reg_state *regs, u32 regno) static void mark_reg_not_init(struct bpf_verifier_env *env,
struct bpf_reg_state *regs, u32 regno)
{ {
if (WARN_ON(regno >= MAX_BPF_REG)) { if (WARN_ON(regno >= MAX_BPF_REG)) {
verbose("mark_reg_not_init(regs, %u)\n", regno); verbose(env, "mark_reg_not_init(regs, %u)\n", regno);
/* Something bad happened, let's kill all regs */ /* Something bad happened, let's kill all regs */
for (regno = 0; regno < MAX_BPF_REG; regno++) for (regno = 0; regno < MAX_BPF_REG; regno++)
__mark_reg_not_init(regs + regno); __mark_reg_not_init(regs + regno);
...@@ -677,22 +678,23 @@ static void mark_reg_not_init(struct bpf_reg_state *regs, u32 regno) ...@@ -677,22 +678,23 @@ static void mark_reg_not_init(struct bpf_reg_state *regs, u32 regno)
__mark_reg_not_init(regs + regno); __mark_reg_not_init(regs + regno);
} }
static void init_reg_state(struct bpf_reg_state *regs) static void init_reg_state(struct bpf_verifier_env *env,
struct bpf_reg_state *regs)
{ {
int i; int i;
for (i = 0; i < MAX_BPF_REG; i++) { for (i = 0; i < MAX_BPF_REG; i++) {
mark_reg_not_init(regs, i); mark_reg_not_init(env, regs, i);
regs[i].live = REG_LIVE_NONE; regs[i].live = REG_LIVE_NONE;
} }
/* frame pointer */ /* frame pointer */
regs[BPF_REG_FP].type = PTR_TO_STACK; regs[BPF_REG_FP].type = PTR_TO_STACK;
mark_reg_known_zero(regs, BPF_REG_FP); mark_reg_known_zero(env, regs, BPF_REG_FP);
/* 1st arg to a function */ /* 1st arg to a function */
regs[BPF_REG_1].type = PTR_TO_CTX; regs[BPF_REG_1].type = PTR_TO_CTX;
mark_reg_known_zero(regs, BPF_REG_1); mark_reg_known_zero(env, regs, BPF_REG_1);
} }
enum reg_arg_type { enum reg_arg_type {
...@@ -726,26 +728,26 @@ static int check_reg_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -726,26 +728,26 @@ static int check_reg_arg(struct bpf_verifier_env *env, u32 regno,
struct bpf_reg_state *regs = env->cur_state.regs; struct bpf_reg_state *regs = env->cur_state.regs;
if (regno >= MAX_BPF_REG) { if (regno >= MAX_BPF_REG) {
verbose("R%d is invalid\n", regno); verbose(env, "R%d is invalid\n", regno);
return -EINVAL; return -EINVAL;
} }
if (t == SRC_OP) { if (t == SRC_OP) {
/* check whether register used as source operand can be read */ /* check whether register used as source operand can be read */
if (regs[regno].type == NOT_INIT) { if (regs[regno].type == NOT_INIT) {
verbose("R%d !read_ok\n", regno); verbose(env, "R%d !read_ok\n", regno);
return -EACCES; return -EACCES;
} }
mark_reg_read(&env->cur_state, regno); mark_reg_read(&env->cur_state, regno);
} else { } else {
/* check whether register used as dest operand can be written to */ /* check whether register used as dest operand can be written to */
if (regno == BPF_REG_FP) { if (regno == BPF_REG_FP) {
verbose("frame pointer is read only\n"); verbose(env, "frame pointer is read only\n");
return -EACCES; return -EACCES;
} }
regs[regno].live |= REG_LIVE_WRITTEN; regs[regno].live |= REG_LIVE_WRITTEN;
if (t == DST_OP) if (t == DST_OP)
mark_reg_unknown(regs, regno); mark_reg_unknown(env, regs, regno);
} }
return 0; return 0;
} }
...@@ -770,7 +772,8 @@ static bool is_spillable_regtype(enum bpf_reg_type type) ...@@ -770,7 +772,8 @@ static bool is_spillable_regtype(enum bpf_reg_type type)
/* check_stack_read/write functions track spill/fill of registers, /* check_stack_read/write functions track spill/fill of registers,
* stack boundary and alignment are checked in check_mem_access() * stack boundary and alignment are checked in check_mem_access()
*/ */
static int check_stack_write(struct bpf_verifier_state *state, int off, static int check_stack_write(struct bpf_verifier_env *env,
struct bpf_verifier_state *state, int off,
int size, int value_regno) int size, int value_regno)
{ {
int i, spi = (MAX_BPF_STACK + off) / BPF_REG_SIZE; int i, spi = (MAX_BPF_STACK + off) / BPF_REG_SIZE;
...@@ -783,7 +786,7 @@ static int check_stack_write(struct bpf_verifier_state *state, int off, ...@@ -783,7 +786,7 @@ static int check_stack_write(struct bpf_verifier_state *state, int off,
/* register containing pointer is being spilled into stack */ /* register containing pointer is being spilled into stack */
if (size != BPF_REG_SIZE) { if (size != BPF_REG_SIZE) {
verbose("invalid size of register spill\n"); verbose(env, "invalid size of register spill\n");
return -EACCES; return -EACCES;
} }
...@@ -818,7 +821,8 @@ static void mark_stack_slot_read(const struct bpf_verifier_state *state, int slo ...@@ -818,7 +821,8 @@ static void mark_stack_slot_read(const struct bpf_verifier_state *state, int slo
} }
} }
static int check_stack_read(struct bpf_verifier_state *state, int off, int size, static int check_stack_read(struct bpf_verifier_env *env,
struct bpf_verifier_state *state, int off, int size,
int value_regno) int value_regno)
{ {
u8 *slot_type; u8 *slot_type;
...@@ -828,12 +832,12 @@ static int check_stack_read(struct bpf_verifier_state *state, int off, int size, ...@@ -828,12 +832,12 @@ static int check_stack_read(struct bpf_verifier_state *state, int off, int size,
if (slot_type[0] == STACK_SPILL) { if (slot_type[0] == STACK_SPILL) {
if (size != BPF_REG_SIZE) { if (size != BPF_REG_SIZE) {
verbose("invalid size of register spill\n"); verbose(env, "invalid size of register spill\n");
return -EACCES; return -EACCES;
} }
for (i = 1; i < BPF_REG_SIZE; i++) { for (i = 1; i < BPF_REG_SIZE; i++) {
if (slot_type[i] != STACK_SPILL) { if (slot_type[i] != STACK_SPILL) {
verbose("corrupted spill memory\n"); verbose(env, "corrupted spill memory\n");
return -EACCES; return -EACCES;
} }
} }
...@@ -849,14 +853,14 @@ static int check_stack_read(struct bpf_verifier_state *state, int off, int size, ...@@ -849,14 +853,14 @@ static int check_stack_read(struct bpf_verifier_state *state, int off, int size,
} else { } else {
for (i = 0; i < size; i++) { for (i = 0; i < size; i++) {
if (slot_type[i] != STACK_MISC) { if (slot_type[i] != STACK_MISC) {
verbose("invalid read from stack off %d+%d size %d\n", verbose(env, "invalid read from stack off %d+%d size %d\n",
off, i, size); off, i, size);
return -EACCES; return -EACCES;
} }
} }
if (value_regno >= 0) if (value_regno >= 0)
/* have read misc data from the stack */ /* have read misc data from the stack */
mark_reg_unknown(state->regs, value_regno); mark_reg_unknown(env, state->regs, value_regno);
return 0; return 0;
} }
} }
...@@ -868,7 +872,7 @@ static int __check_map_access(struct bpf_verifier_env *env, u32 regno, int off, ...@@ -868,7 +872,7 @@ static int __check_map_access(struct bpf_verifier_env *env, u32 regno, int off,
struct bpf_map *map = env->cur_state.regs[regno].map_ptr; struct bpf_map *map = env->cur_state.regs[regno].map_ptr;
if (off < 0 || size <= 0 || off + size > map->value_size) { if (off < 0 || size <= 0 || off + size > map->value_size) {
verbose("invalid access to map value, value_size=%d off=%d size=%d\n", verbose(env, "invalid access to map value, value_size=%d off=%d size=%d\n",
map->value_size, off, size); map->value_size, off, size);
return -EACCES; return -EACCES;
} }
...@@ -887,8 +891,8 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno, ...@@ -887,8 +891,8 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno,
* need to try adding each of min_value and max_value to off * need to try adding each of min_value and max_value to off
* to make sure our theoretical access will be safe. * to make sure our theoretical access will be safe.
*/ */
if (verifier_log.level) if (env->log.level)
print_verifier_state(state); print_verifier_state(env, state);
/* The minimum value is only important with signed /* The minimum value is only important with signed
* comparisons where we can't assume the floor of a * comparisons where we can't assume the floor of a
* value is 0. If we are using signed variables for our * value is 0. If we are using signed variables for our
...@@ -896,13 +900,14 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno, ...@@ -896,13 +900,14 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno,
* will have a set floor within our range. * will have a set floor within our range.
*/ */
if (reg->smin_value < 0) { if (reg->smin_value < 0) {
verbose("R%d min value is negative, either use unsigned index or do a if (index >=0) check.\n", verbose(env, "R%d min value is negative, either use unsigned index or do a if (index >=0) check.\n",
regno); regno);
return -EACCES; return -EACCES;
} }
err = __check_map_access(env, regno, reg->smin_value + off, size); err = __check_map_access(env, regno, reg->smin_value + off, size);
if (err) { if (err) {
verbose("R%d min value is outside of the array range\n", regno); verbose(env, "R%d min value is outside of the array range\n",
regno);
return err; return err;
} }
...@@ -911,13 +916,14 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno, ...@@ -911,13 +916,14 @@ static int check_map_access(struct bpf_verifier_env *env, u32 regno,
* If reg->umax_value + off could overflow, treat that as unbounded too. * If reg->umax_value + off could overflow, treat that as unbounded too.
*/ */
if (reg->umax_value >= BPF_MAX_VAR_OFF) { if (reg->umax_value >= BPF_MAX_VAR_OFF) {
verbose("R%d unbounded memory access, make sure to bounds check any array access into a map\n", verbose(env, "R%d unbounded memory access, make sure to bounds check any array access into a map\n",
regno); regno);
return -EACCES; return -EACCES;
} }
err = __check_map_access(env, regno, reg->umax_value + off, size); err = __check_map_access(env, regno, reg->umax_value + off, size);
if (err) if (err)
verbose("R%d max value is outside of the array range\n", regno); verbose(env, "R%d max value is outside of the array range\n",
regno);
return err; return err;
} }
...@@ -956,7 +962,7 @@ static int __check_packet_access(struct bpf_verifier_env *env, u32 regno, ...@@ -956,7 +962,7 @@ static int __check_packet_access(struct bpf_verifier_env *env, u32 regno,
struct bpf_reg_state *reg = &regs[regno]; struct bpf_reg_state *reg = &regs[regno];
if (off < 0 || size <= 0 || (u64)off + size > reg->range) { if (off < 0 || size <= 0 || (u64)off + size > reg->range) {
verbose("invalid access to packet, off=%d size=%d, R%d(id=%d,off=%d,r=%d)\n", verbose(env, "invalid access to packet, off=%d size=%d, R%d(id=%d,off=%d,r=%d)\n",
off, size, regno, reg->id, reg->off, reg->range); off, size, regno, reg->id, reg->off, reg->range);
return -EACCES; return -EACCES;
} }
...@@ -979,13 +985,13 @@ static int check_packet_access(struct bpf_verifier_env *env, u32 regno, int off, ...@@ -979,13 +985,13 @@ static int check_packet_access(struct bpf_verifier_env *env, u32 regno, int off,
* detail to prove they're safe. * detail to prove they're safe.
*/ */
if (reg->smin_value < 0) { if (reg->smin_value < 0) {
verbose("R%d min value is negative, either use unsigned index or do a if (index >=0) check.\n", verbose(env, "R%d min value is negative, either use unsigned index or do a if (index >=0) check.\n",
regno); regno);
return -EACCES; return -EACCES;
} }
err = __check_packet_access(env, regno, off, size); err = __check_packet_access(env, regno, off, size);
if (err) { if (err) {
verbose("R%d offset is outside of the packet\n", regno); verbose(env, "R%d offset is outside of the packet\n", regno);
return err; return err;
} }
return err; return err;
...@@ -1021,7 +1027,7 @@ static int check_ctx_access(struct bpf_verifier_env *env, int insn_idx, int off, ...@@ -1021,7 +1027,7 @@ static int check_ctx_access(struct bpf_verifier_env *env, int insn_idx, int off,
return 0; return 0;
} }
verbose("invalid bpf_context access off=%d size=%d\n", off, size); verbose(env, "invalid bpf_context access off=%d size=%d\n", off, size);
return -EACCES; return -EACCES;
} }
...@@ -1039,7 +1045,8 @@ static bool is_pointer_value(struct bpf_verifier_env *env, int regno) ...@@ -1039,7 +1045,8 @@ static bool is_pointer_value(struct bpf_verifier_env *env, int regno)
return __is_pointer_value(env->allow_ptr_leaks, &env->cur_state.regs[regno]); return __is_pointer_value(env->allow_ptr_leaks, &env->cur_state.regs[regno]);
} }
static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, static int check_pkt_ptr_alignment(struct bpf_verifier_env *env,
const struct bpf_reg_state *reg,
int off, int size, bool strict) int off, int size, bool strict)
{ {
struct tnum reg_off; struct tnum reg_off;
...@@ -1064,7 +1071,8 @@ static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, ...@@ -1064,7 +1071,8 @@ static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg,
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose("misaligned packet access off %d+%s+%d+%d size %d\n", verbose(env,
"misaligned packet access off %d+%s+%d+%d size %d\n",
ip_align, tn_buf, reg->off, off, size); ip_align, tn_buf, reg->off, off, size);
return -EACCES; return -EACCES;
} }
...@@ -1072,7 +1080,8 @@ static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, ...@@ -1072,7 +1080,8 @@ static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg,
return 0; return 0;
} }
static int check_generic_ptr_alignment(const struct bpf_reg_state *reg, static int check_generic_ptr_alignment(struct bpf_verifier_env *env,
const struct bpf_reg_state *reg,
const char *pointer_desc, const char *pointer_desc,
int off, int size, bool strict) int off, int size, bool strict)
{ {
...@@ -1087,7 +1096,7 @@ static int check_generic_ptr_alignment(const struct bpf_reg_state *reg, ...@@ -1087,7 +1096,7 @@ static int check_generic_ptr_alignment(const struct bpf_reg_state *reg,
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose("misaligned %saccess off %s+%d+%d size %d\n", verbose(env, "misaligned %saccess off %s+%d+%d size %d\n",
pointer_desc, tn_buf, reg->off, off, size); pointer_desc, tn_buf, reg->off, off, size);
return -EACCES; return -EACCES;
} }
...@@ -1108,7 +1117,7 @@ static int check_ptr_alignment(struct bpf_verifier_env *env, ...@@ -1108,7 +1117,7 @@ static int check_ptr_alignment(struct bpf_verifier_env *env,
/* Special case, because of NET_IP_ALIGN. Given metadata sits /* Special case, because of NET_IP_ALIGN. Given metadata sits
* right in front, treat it the very same way. * right in front, treat it the very same way.
*/ */
return check_pkt_ptr_alignment(reg, off, size, strict); return check_pkt_ptr_alignment(env, reg, off, size, strict);
case PTR_TO_MAP_VALUE: case PTR_TO_MAP_VALUE:
pointer_desc = "value "; pointer_desc = "value ";
break; break;
...@@ -1121,7 +1130,8 @@ static int check_ptr_alignment(struct bpf_verifier_env *env, ...@@ -1121,7 +1130,8 @@ static int check_ptr_alignment(struct bpf_verifier_env *env,
default: default:
break; break;
} }
return check_generic_ptr_alignment(reg, pointer_desc, off, size, strict); return check_generic_ptr_alignment(env, reg, pointer_desc, off, size,
strict);
} }
/* check whether memory at (regno + off) is accessible for t = (read | write) /* check whether memory at (regno + off) is accessible for t = (read | write)
...@@ -1153,20 +1163,20 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn ...@@ -1153,20 +1163,20 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn
if (reg->type == PTR_TO_MAP_VALUE) { if (reg->type == PTR_TO_MAP_VALUE) {
if (t == BPF_WRITE && value_regno >= 0 && if (t == BPF_WRITE && value_regno >= 0 &&
is_pointer_value(env, value_regno)) { is_pointer_value(env, value_regno)) {
verbose("R%d leaks addr into map\n", value_regno); verbose(env, "R%d leaks addr into map\n", value_regno);
return -EACCES; return -EACCES;
} }
err = check_map_access(env, regno, off, size); err = check_map_access(env, regno, off, size);
if (!err && t == BPF_READ && value_regno >= 0) if (!err && t == BPF_READ && value_regno >= 0)
mark_reg_unknown(state->regs, value_regno); mark_reg_unknown(env, state->regs, value_regno);
} else if (reg->type == PTR_TO_CTX) { } else if (reg->type == PTR_TO_CTX) {
enum bpf_reg_type reg_type = SCALAR_VALUE; enum bpf_reg_type reg_type = SCALAR_VALUE;
if (t == BPF_WRITE && value_regno >= 0 && if (t == BPF_WRITE && value_regno >= 0 &&
is_pointer_value(env, value_regno)) { is_pointer_value(env, value_regno)) {
verbose("R%d leaks addr into ctx\n", value_regno); verbose(env, "R%d leaks addr into ctx\n", value_regno);
return -EACCES; return -EACCES;
} }
/* ctx accesses must be at a fixed offset, so that we can /* ctx accesses must be at a fixed offset, so that we can
...@@ -1176,7 +1186,8 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn ...@@ -1176,7 +1186,8 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose("variable ctx access var_off=%s off=%d size=%d", verbose(env,
"variable ctx access var_off=%s off=%d size=%d",
tn_buf, off, size); tn_buf, off, size);
return -EACCES; return -EACCES;
} }
...@@ -1188,9 +1199,10 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn ...@@ -1188,9 +1199,10 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn
* case, we know the offset is zero. * case, we know the offset is zero.
*/ */
if (reg_type == SCALAR_VALUE) if (reg_type == SCALAR_VALUE)
mark_reg_unknown(state->regs, value_regno); mark_reg_unknown(env, state->regs, value_regno);
else else
mark_reg_known_zero(state->regs, value_regno); mark_reg_known_zero(env, state->regs,
value_regno);
state->regs[value_regno].id = 0; state->regs[value_regno].id = 0;
state->regs[value_regno].off = 0; state->regs[value_regno].off = 0;
state->regs[value_regno].range = 0; state->regs[value_regno].range = 0;
...@@ -1206,13 +1218,14 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn ...@@ -1206,13 +1218,14 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose("variable stack access var_off=%s off=%d size=%d", verbose(env, "variable stack access var_off=%s off=%d size=%d",
tn_buf, off, size); tn_buf, off, size);
return -EACCES; return -EACCES;
} }
off += reg->var_off.value; off += reg->var_off.value;
if (off >= 0 || off < -MAX_BPF_STACK) { if (off >= 0 || off < -MAX_BPF_STACK) {
verbose("invalid stack off=%d size=%d\n", off, size); verbose(env, "invalid stack off=%d size=%d\n", off,
size);
return -EACCES; return -EACCES;
} }
...@@ -1223,29 +1236,32 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn ...@@ -1223,29 +1236,32 @@ static int check_mem_access(struct bpf_verifier_env *env, int insn_idx, u32 regn
if (!env->allow_ptr_leaks && if (!env->allow_ptr_leaks &&
state->stack_slot_type[MAX_BPF_STACK + off] == STACK_SPILL && state->stack_slot_type[MAX_BPF_STACK + off] == STACK_SPILL &&
size != BPF_REG_SIZE) { size != BPF_REG_SIZE) {
verbose("attempt to corrupt spilled pointer on stack\n"); verbose(env, "attempt to corrupt spilled pointer on stack\n");
return -EACCES; return -EACCES;
} }
err = check_stack_write(state, off, size, value_regno); err = check_stack_write(env, state, off, size,
value_regno);
} else { } else {
err = check_stack_read(state, off, size, value_regno); err = check_stack_read(env, state, off, size,
value_regno);
} }
} else if (reg_is_pkt_pointer(reg)) { } else if (reg_is_pkt_pointer(reg)) {
if (t == BPF_WRITE && !may_access_direct_pkt_data(env, NULL, t)) { if (t == BPF_WRITE && !may_access_direct_pkt_data(env, NULL, t)) {
verbose("cannot write into packet\n"); verbose(env, "cannot write into packet\n");
return -EACCES; return -EACCES;
} }
if (t == BPF_WRITE && value_regno >= 0 && if (t == BPF_WRITE && value_regno >= 0 &&
is_pointer_value(env, value_regno)) { is_pointer_value(env, value_regno)) {
verbose("R%d leaks addr into packet\n", value_regno); verbose(env, "R%d leaks addr into packet\n",
value_regno);
return -EACCES; return -EACCES;
} }
err = check_packet_access(env, regno, off, size); err = check_packet_access(env, regno, off, size);
if (!err && t == BPF_READ && value_regno >= 0) if (!err && t == BPF_READ && value_regno >= 0)
mark_reg_unknown(state->regs, value_regno); mark_reg_unknown(env, state->regs, value_regno);
} else { } else {
verbose("R%d invalid mem access '%s'\n", verbose(env, "R%d invalid mem access '%s'\n", regno,
regno, reg_type_str[reg->type]); reg_type_str[reg->type]);
return -EACCES; return -EACCES;
} }
...@@ -1265,7 +1281,7 @@ static int check_xadd(struct bpf_verifier_env *env, int insn_idx, struct bpf_ins ...@@ -1265,7 +1281,7 @@ static int check_xadd(struct bpf_verifier_env *env, int insn_idx, struct bpf_ins
if ((BPF_SIZE(insn->code) != BPF_W && BPF_SIZE(insn->code) != BPF_DW) || if ((BPF_SIZE(insn->code) != BPF_W && BPF_SIZE(insn->code) != BPF_DW) ||
insn->imm != 0) { insn->imm != 0) {
verbose("BPF_XADD uses reserved fields\n"); verbose(env, "BPF_XADD uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -1280,7 +1296,7 @@ static int check_xadd(struct bpf_verifier_env *env, int insn_idx, struct bpf_ins ...@@ -1280,7 +1296,7 @@ static int check_xadd(struct bpf_verifier_env *env, int insn_idx, struct bpf_ins
return err; return err;
if (is_pointer_value(env, insn->src_reg)) { if (is_pointer_value(env, insn->src_reg)) {
verbose("R%d leaks addr into mem\n", insn->src_reg); verbose(env, "R%d leaks addr into mem\n", insn->src_reg);
return -EACCES; return -EACCES;
} }
...@@ -1321,7 +1337,7 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno, ...@@ -1321,7 +1337,7 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno,
register_is_null(regs[regno])) register_is_null(regs[regno]))
return 0; return 0;
verbose("R%d type=%s expected=%s\n", regno, verbose(env, "R%d type=%s expected=%s\n", regno,
reg_type_str[regs[regno].type], reg_type_str[regs[regno].type],
reg_type_str[PTR_TO_STACK]); reg_type_str[PTR_TO_STACK]);
return -EACCES; return -EACCES;
...@@ -1332,13 +1348,13 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno, ...@@ -1332,13 +1348,13 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno,
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), regs[regno].var_off); tnum_strn(tn_buf, sizeof(tn_buf), regs[regno].var_off);
verbose("invalid variable stack read R%d var_off=%s\n", verbose(env, "invalid variable stack read R%d var_off=%s\n",
regno, tn_buf); regno, tn_buf);
} }
off = regs[regno].off + regs[regno].var_off.value; off = regs[regno].off + regs[regno].var_off.value;
if (off >= 0 || off < -MAX_BPF_STACK || off + access_size > 0 || if (off >= 0 || off < -MAX_BPF_STACK || off + access_size > 0 ||
access_size <= 0) { access_size <= 0) {
verbose("invalid stack type R%d off=%d access_size=%d\n", verbose(env, "invalid stack type R%d off=%d access_size=%d\n",
regno, off, access_size); regno, off, access_size);
return -EACCES; return -EACCES;
} }
...@@ -1354,7 +1370,7 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno, ...@@ -1354,7 +1370,7 @@ static int check_stack_boundary(struct bpf_verifier_env *env, int regno,
for (i = 0; i < access_size; i++) { for (i = 0; i < access_size; i++) {
if (state->stack_slot_type[MAX_BPF_STACK + off + i] != STACK_MISC) { if (state->stack_slot_type[MAX_BPF_STACK + off + i] != STACK_MISC) {
verbose("invalid indirect read from stack off %d+%d size %d\n", verbose(env, "invalid indirect read from stack off %d+%d size %d\n",
off, i, access_size); off, i, access_size);
return -EACCES; return -EACCES;
} }
...@@ -1397,7 +1413,8 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1397,7 +1413,8 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
if (arg_type == ARG_ANYTHING) { if (arg_type == ARG_ANYTHING) {
if (is_pointer_value(env, regno)) { if (is_pointer_value(env, regno)) {
verbose("R%d leaks addr into helper function\n", regno); verbose(env, "R%d leaks addr into helper function\n",
regno);
return -EACCES; return -EACCES;
} }
return 0; return 0;
...@@ -1405,7 +1422,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1405,7 +1422,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
if (type_is_pkt_pointer(type) && if (type_is_pkt_pointer(type) &&
!may_access_direct_pkt_data(env, meta, BPF_READ)) { !may_access_direct_pkt_data(env, meta, BPF_READ)) {
verbose("helper access to the packet is not allowed\n"); verbose(env, "helper access to the packet is not allowed\n");
return -EACCES; return -EACCES;
} }
...@@ -1443,7 +1460,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1443,7 +1460,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
goto err_type; goto err_type;
meta->raw_mode = arg_type == ARG_PTR_TO_UNINIT_MEM; meta->raw_mode = arg_type == ARG_PTR_TO_UNINIT_MEM;
} else { } else {
verbose("unsupported arg_type %d\n", arg_type); verbose(env, "unsupported arg_type %d\n", arg_type);
return -EFAULT; return -EFAULT;
} }
...@@ -1461,7 +1478,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1461,7 +1478,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
* we have to check map_key here. Otherwise it means * we have to check map_key here. Otherwise it means
* that kernel subsystem misconfigured verifier * that kernel subsystem misconfigured verifier
*/ */
verbose("invalid map_ptr to access map->key\n"); verbose(env, "invalid map_ptr to access map->key\n");
return -EACCES; return -EACCES;
} }
if (type_is_pkt_pointer(type)) if (type_is_pkt_pointer(type))
...@@ -1477,7 +1494,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1477,7 +1494,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
*/ */
if (!meta->map_ptr) { if (!meta->map_ptr) {
/* kernel subsystem misconfigured verifier */ /* kernel subsystem misconfigured verifier */
verbose("invalid map_ptr to access map->value\n"); verbose(env, "invalid map_ptr to access map->value\n");
return -EACCES; return -EACCES;
} }
if (type_is_pkt_pointer(type)) if (type_is_pkt_pointer(type))
...@@ -1497,7 +1514,8 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1497,7 +1514,8 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
*/ */
if (regno == 0) { if (regno == 0) {
/* kernel subsystem misconfigured verifier */ /* kernel subsystem misconfigured verifier */
verbose("ARG_CONST_SIZE cannot be first argument\n"); verbose(env,
"ARG_CONST_SIZE cannot be first argument\n");
return -EACCES; return -EACCES;
} }
...@@ -1514,7 +1532,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1514,7 +1532,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
meta = NULL; meta = NULL;
if (reg->smin_value < 0) { if (reg->smin_value < 0) {
verbose("R%d min value is negative, either use unsigned or 'var &= const'\n", verbose(env, "R%d min value is negative, either use unsigned or 'var &= const'\n",
regno); regno);
return -EACCES; return -EACCES;
} }
...@@ -1528,7 +1546,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1528,7 +1546,7 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
} }
if (reg->umax_value >= BPF_MAX_VAR_SIZ) { if (reg->umax_value >= BPF_MAX_VAR_SIZ) {
verbose("R%d unbounded memory access, use 'var &= const' or 'if (var < const)'\n", verbose(env, "R%d unbounded memory access, use 'var &= const' or 'if (var < const)'\n",
regno); regno);
return -EACCES; return -EACCES;
} }
...@@ -1539,12 +1557,13 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno, ...@@ -1539,12 +1557,13 @@ static int check_func_arg(struct bpf_verifier_env *env, u32 regno,
return err; return err;
err_type: err_type:
verbose("R%d type=%s expected=%s\n", regno, verbose(env, "R%d type=%s expected=%s\n", regno,
reg_type_str[type], reg_type_str[expected_type]); reg_type_str[type], reg_type_str[expected_type]);
return -EACCES; return -EACCES;
} }
static int check_map_func_compatibility(struct bpf_map *map, int func_id) static int check_map_func_compatibility(struct bpf_verifier_env *env,
struct bpf_map *map, int func_id)
{ {
if (!map) if (!map)
return 0; return 0;
...@@ -1632,7 +1651,7 @@ static int check_map_func_compatibility(struct bpf_map *map, int func_id) ...@@ -1632,7 +1651,7 @@ static int check_map_func_compatibility(struct bpf_map *map, int func_id)
return 0; return 0;
error: error:
verbose("cannot pass map_type %d into func %s#%d\n", verbose(env, "cannot pass map_type %d into func %s#%d\n",
map->map_type, func_id_name(func_id), func_id); map->map_type, func_id_name(func_id), func_id);
return -EINVAL; return -EINVAL;
} }
...@@ -1666,7 +1685,7 @@ static void clear_all_pkt_pointers(struct bpf_verifier_env *env) ...@@ -1666,7 +1685,7 @@ static void clear_all_pkt_pointers(struct bpf_verifier_env *env)
for (i = 0; i < MAX_BPF_REG; i++) for (i = 0; i < MAX_BPF_REG; i++)
if (reg_is_pkt_pointer_any(&regs[i])) if (reg_is_pkt_pointer_any(&regs[i]))
mark_reg_unknown(regs, i); mark_reg_unknown(env, regs, i);
for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) { for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) {
if (state->stack_slot_type[i] != STACK_SPILL) if (state->stack_slot_type[i] != STACK_SPILL)
...@@ -1688,7 +1707,8 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1688,7 +1707,8 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
/* find function prototype */ /* find function prototype */
if (func_id < 0 || func_id >= __BPF_FUNC_MAX_ID) { if (func_id < 0 || func_id >= __BPF_FUNC_MAX_ID) {
verbose("invalid func %s#%d\n", func_id_name(func_id), func_id); verbose(env, "invalid func %s#%d\n", func_id_name(func_id),
func_id);
return -EINVAL; return -EINVAL;
} }
...@@ -1696,13 +1716,14 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1696,13 +1716,14 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
fn = env->prog->aux->ops->get_func_proto(func_id); fn = env->prog->aux->ops->get_func_proto(func_id);
if (!fn) { if (!fn) {
verbose("unknown func %s#%d\n", func_id_name(func_id), func_id); verbose(env, "unknown func %s#%d\n", func_id_name(func_id),
func_id);
return -EINVAL; return -EINVAL;
} }
/* eBPF programs must be GPL compatible to use GPL-ed functions */ /* eBPF programs must be GPL compatible to use GPL-ed functions */
if (!env->prog->gpl_compatible && fn->gpl_only) { if (!env->prog->gpl_compatible && fn->gpl_only) {
verbose("cannot call GPL only function from proprietary program\n"); verbose(env, "cannot call GPL only function from proprietary program\n");
return -EINVAL; return -EINVAL;
} }
...@@ -1716,7 +1737,7 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1716,7 +1737,7 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
*/ */
err = check_raw_mode(fn); err = check_raw_mode(fn);
if (err) { if (err) {
verbose("kernel subsystem misconfigured func %s#%d\n", verbose(env, "kernel subsystem misconfigured func %s#%d\n",
func_id_name(func_id), func_id); func_id_name(func_id), func_id);
return err; return err;
} }
...@@ -1749,14 +1770,14 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1749,14 +1770,14 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
/* reset caller saved regs */ /* reset caller saved regs */
for (i = 0; i < CALLER_SAVED_REGS; i++) { for (i = 0; i < CALLER_SAVED_REGS; i++) {
mark_reg_not_init(regs, caller_saved[i]); mark_reg_not_init(env, regs, caller_saved[i]);
check_reg_arg(env, caller_saved[i], DST_OP_NO_MARK); check_reg_arg(env, caller_saved[i], DST_OP_NO_MARK);
} }
/* update return register (already marked as written above) */ /* update return register (already marked as written above) */
if (fn->ret_type == RET_INTEGER) { if (fn->ret_type == RET_INTEGER) {
/* sets type to SCALAR_VALUE */ /* sets type to SCALAR_VALUE */
mark_reg_unknown(regs, BPF_REG_0); mark_reg_unknown(env, regs, BPF_REG_0);
} else if (fn->ret_type == RET_VOID) { } else if (fn->ret_type == RET_VOID) {
regs[BPF_REG_0].type = NOT_INIT; regs[BPF_REG_0].type = NOT_INIT;
} else if (fn->ret_type == RET_PTR_TO_MAP_VALUE_OR_NULL) { } else if (fn->ret_type == RET_PTR_TO_MAP_VALUE_OR_NULL) {
...@@ -1764,14 +1785,15 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1764,14 +1785,15 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
regs[BPF_REG_0].type = PTR_TO_MAP_VALUE_OR_NULL; regs[BPF_REG_0].type = PTR_TO_MAP_VALUE_OR_NULL;
/* There is no offset yet applied, variable or fixed */ /* There is no offset yet applied, variable or fixed */
mark_reg_known_zero(regs, BPF_REG_0); mark_reg_known_zero(env, regs, BPF_REG_0);
regs[BPF_REG_0].off = 0; regs[BPF_REG_0].off = 0;
/* remember map_ptr, so that check_map_access() /* remember map_ptr, so that check_map_access()
* can check 'value_size' boundary of memory access * can check 'value_size' boundary of memory access
* to map element returned from bpf_map_lookup_elem() * to map element returned from bpf_map_lookup_elem()
*/ */
if (meta.map_ptr == NULL) { if (meta.map_ptr == NULL) {
verbose("kernel subsystem misconfigured verifier\n"); verbose(env,
"kernel subsystem misconfigured verifier\n");
return -EINVAL; return -EINVAL;
} }
regs[BPF_REG_0].map_ptr = meta.map_ptr; regs[BPF_REG_0].map_ptr = meta.map_ptr;
...@@ -1782,12 +1804,12 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx) ...@@ -1782,12 +1804,12 @@ static int check_call(struct bpf_verifier_env *env, int func_id, int insn_idx)
else if (insn_aux->map_ptr != meta.map_ptr) else if (insn_aux->map_ptr != meta.map_ptr)
insn_aux->map_ptr = BPF_MAP_PTR_POISON; insn_aux->map_ptr = BPF_MAP_PTR_POISON;
} else { } else {
verbose("unknown return type %d of func %s#%d\n", verbose(env, "unknown return type %d of func %s#%d\n",
fn->ret_type, func_id_name(func_id), func_id); fn->ret_type, func_id_name(func_id), func_id);
return -EINVAL; return -EINVAL;
} }
err = check_map_func_compatibility(meta.map_ptr, func_id); err = check_map_func_compatibility(env, meta.map_ptr, func_id);
if (err) if (err)
return err; return err;
...@@ -1846,39 +1868,42 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env, ...@@ -1846,39 +1868,42 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env,
dst_reg = &regs[dst]; dst_reg = &regs[dst];
if (WARN_ON_ONCE(known && (smin_val != smax_val))) { if (WARN_ON_ONCE(known && (smin_val != smax_val))) {
print_verifier_state(&env->cur_state); print_verifier_state(env, &env->cur_state);
verbose("verifier internal error: known but bad sbounds\n"); verbose(env,
"verifier internal error: known but bad sbounds\n");
return -EINVAL; return -EINVAL;
} }
if (WARN_ON_ONCE(known && (umin_val != umax_val))) { if (WARN_ON_ONCE(known && (umin_val != umax_val))) {
print_verifier_state(&env->cur_state); print_verifier_state(env, &env->cur_state);
verbose("verifier internal error: known but bad ubounds\n"); verbose(env,
"verifier internal error: known but bad ubounds\n");
return -EINVAL; return -EINVAL;
} }
if (BPF_CLASS(insn->code) != BPF_ALU64) { if (BPF_CLASS(insn->code) != BPF_ALU64) {
/* 32-bit ALU ops on pointers produce (meaningless) scalars */ /* 32-bit ALU ops on pointers produce (meaningless) scalars */
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d 32-bit pointer arithmetic prohibited\n", verbose(env,
"R%d 32-bit pointer arithmetic prohibited\n",
dst); dst);
return -EACCES; return -EACCES;
} }
if (ptr_reg->type == PTR_TO_MAP_VALUE_OR_NULL) { if (ptr_reg->type == PTR_TO_MAP_VALUE_OR_NULL) {
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d pointer arithmetic on PTR_TO_MAP_VALUE_OR_NULL prohibited, null-check it first\n", verbose(env, "R%d pointer arithmetic on PTR_TO_MAP_VALUE_OR_NULL prohibited, null-check it first\n",
dst); dst);
return -EACCES; return -EACCES;
} }
if (ptr_reg->type == CONST_PTR_TO_MAP) { if (ptr_reg->type == CONST_PTR_TO_MAP) {
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d pointer arithmetic on CONST_PTR_TO_MAP prohibited\n", verbose(env, "R%d pointer arithmetic on CONST_PTR_TO_MAP prohibited\n",
dst); dst);
return -EACCES; return -EACCES;
} }
if (ptr_reg->type == PTR_TO_PACKET_END) { if (ptr_reg->type == PTR_TO_PACKET_END) {
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d pointer arithmetic on PTR_TO_PACKET_END prohibited\n", verbose(env, "R%d pointer arithmetic on PTR_TO_PACKET_END prohibited\n",
dst); dst);
return -EACCES; return -EACCES;
} }
...@@ -1943,7 +1968,7 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env, ...@@ -1943,7 +1968,7 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env,
if (dst_reg == off_reg) { if (dst_reg == off_reg) {
/* scalar -= pointer. Creates an unknown scalar */ /* scalar -= pointer. Creates an unknown scalar */
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d tried to subtract pointer from scalar\n", verbose(env, "R%d tried to subtract pointer from scalar\n",
dst); dst);
return -EACCES; return -EACCES;
} }
...@@ -1953,7 +1978,7 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env, ...@@ -1953,7 +1978,7 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env,
*/ */
if (ptr_reg->type == PTR_TO_STACK) { if (ptr_reg->type == PTR_TO_STACK) {
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d subtraction from stack pointer prohibited\n", verbose(env, "R%d subtraction from stack pointer prohibited\n",
dst); dst);
return -EACCES; return -EACCES;
} }
...@@ -2008,13 +2033,13 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env, ...@@ -2008,13 +2033,13 @@ static int adjust_ptr_min_max_vals(struct bpf_verifier_env *env,
* ptr &= ~3 which would reduce min_value by 3.) * ptr &= ~3 which would reduce min_value by 3.)
*/ */
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d bitwise operator %s on pointer prohibited\n", verbose(env, "R%d bitwise operator %s on pointer prohibited\n",
dst, bpf_alu_string[opcode >> 4]); dst, bpf_alu_string[opcode >> 4]);
return -EACCES; return -EACCES;
default: default:
/* other operators (e.g. MUL,LSH) produce non-pointer results */ /* other operators (e.g. MUL,LSH) produce non-pointer results */
if (!env->allow_ptr_leaks) if (!env->allow_ptr_leaks)
verbose("R%d pointer arithmetic with %s operator prohibited\n", verbose(env, "R%d pointer arithmetic with %s operator prohibited\n",
dst, bpf_alu_string[opcode >> 4]); dst, bpf_alu_string[opcode >> 4]);
return -EACCES; return -EACCES;
} }
...@@ -2180,7 +2205,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env, ...@@ -2180,7 +2205,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env,
/* Shifts greater than 63 are undefined. This includes /* Shifts greater than 63 are undefined. This includes
* shifts by a negative number. * shifts by a negative number.
*/ */
mark_reg_unknown(regs, insn->dst_reg); mark_reg_unknown(env, regs, insn->dst_reg);
break; break;
} }
/* We lose all sign bit information (except what we can pick /* We lose all sign bit information (except what we can pick
...@@ -2208,7 +2233,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env, ...@@ -2208,7 +2233,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env,
/* Shifts greater than 63 are undefined. This includes /* Shifts greater than 63 are undefined. This includes
* shifts by a negative number. * shifts by a negative number.
*/ */
mark_reg_unknown(regs, insn->dst_reg); mark_reg_unknown(env, regs, insn->dst_reg);
break; break;
} }
/* BPF_RSH is an unsigned shift, so make the appropriate casts */ /* BPF_RSH is an unsigned shift, so make the appropriate casts */
...@@ -2236,7 +2261,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env, ...@@ -2236,7 +2261,7 @@ static int adjust_scalar_min_max_vals(struct bpf_verifier_env *env,
__update_reg_bounds(dst_reg); __update_reg_bounds(dst_reg);
break; break;
default: default:
mark_reg_unknown(regs, insn->dst_reg); mark_reg_unknown(env, regs, insn->dst_reg);
break; break;
} }
...@@ -2268,12 +2293,12 @@ static int adjust_reg_min_max_vals(struct bpf_verifier_env *env, ...@@ -2268,12 +2293,12 @@ static int adjust_reg_min_max_vals(struct bpf_verifier_env *env,
* an arbitrary scalar. * an arbitrary scalar.
*/ */
if (!env->allow_ptr_leaks) { if (!env->allow_ptr_leaks) {
verbose("R%d pointer %s pointer prohibited\n", verbose(env, "R%d pointer %s pointer prohibited\n",
insn->dst_reg, insn->dst_reg,
bpf_alu_string[opcode >> 4]); bpf_alu_string[opcode >> 4]);
return -EACCES; return -EACCES;
} }
mark_reg_unknown(regs, insn->dst_reg); mark_reg_unknown(env, regs, insn->dst_reg);
return 0; return 0;
} else { } else {
/* scalar += pointer /* scalar += pointer
...@@ -2325,13 +2350,13 @@ static int adjust_reg_min_max_vals(struct bpf_verifier_env *env, ...@@ -2325,13 +2350,13 @@ static int adjust_reg_min_max_vals(struct bpf_verifier_env *env,
/* Got here implies adding two SCALAR_VALUEs */ /* Got here implies adding two SCALAR_VALUEs */
if (WARN_ON_ONCE(ptr_reg)) { if (WARN_ON_ONCE(ptr_reg)) {
print_verifier_state(&env->cur_state); print_verifier_state(env, &env->cur_state);
verbose("verifier internal error: unexpected ptr_reg\n"); verbose(env, "verifier internal error: unexpected ptr_reg\n");
return -EINVAL; return -EINVAL;
} }
if (WARN_ON(!src_reg)) { if (WARN_ON(!src_reg)) {
print_verifier_state(&env->cur_state); print_verifier_state(env, &env->cur_state);
verbose("verifier internal error: no src_reg\n"); verbose(env, "verifier internal error: no src_reg\n");
return -EINVAL; return -EINVAL;
} }
return adjust_scalar_min_max_vals(env, insn, dst_reg, *src_reg); return adjust_scalar_min_max_vals(env, insn, dst_reg, *src_reg);
...@@ -2349,14 +2374,14 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2349,14 +2374,14 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
if (BPF_SRC(insn->code) != 0 || if (BPF_SRC(insn->code) != 0 ||
insn->src_reg != BPF_REG_0 || insn->src_reg != BPF_REG_0 ||
insn->off != 0 || insn->imm != 0) { insn->off != 0 || insn->imm != 0) {
verbose("BPF_NEG uses reserved fields\n"); verbose(env, "BPF_NEG uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
} else { } else {
if (insn->src_reg != BPF_REG_0 || insn->off != 0 || if (insn->src_reg != BPF_REG_0 || insn->off != 0 ||
(insn->imm != 16 && insn->imm != 32 && insn->imm != 64) || (insn->imm != 16 && insn->imm != 32 && insn->imm != 64) ||
BPF_CLASS(insn->code) == BPF_ALU64) { BPF_CLASS(insn->code) == BPF_ALU64) {
verbose("BPF_END uses reserved fields\n"); verbose(env, "BPF_END uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
} }
...@@ -2367,7 +2392,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2367,7 +2392,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
return err; return err;
if (is_pointer_value(env, insn->dst_reg)) { if (is_pointer_value(env, insn->dst_reg)) {
verbose("R%d pointer arithmetic prohibited\n", verbose(env, "R%d pointer arithmetic prohibited\n",
insn->dst_reg); insn->dst_reg);
return -EACCES; return -EACCES;
} }
...@@ -2381,7 +2406,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2381,7 +2406,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
if (BPF_SRC(insn->code) == BPF_X) { if (BPF_SRC(insn->code) == BPF_X) {
if (insn->imm != 0 || insn->off != 0) { if (insn->imm != 0 || insn->off != 0) {
verbose("BPF_MOV uses reserved fields\n"); verbose(env, "BPF_MOV uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -2391,7 +2416,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2391,7 +2416,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
return err; return err;
} else { } else {
if (insn->src_reg != BPF_REG_0 || insn->off != 0) { if (insn->src_reg != BPF_REG_0 || insn->off != 0) {
verbose("BPF_MOV uses reserved fields\n"); verbose(env, "BPF_MOV uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
} }
...@@ -2411,11 +2436,12 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2411,11 +2436,12 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
} else { } else {
/* R1 = (u32) R2 */ /* R1 = (u32) R2 */
if (is_pointer_value(env, insn->src_reg)) { if (is_pointer_value(env, insn->src_reg)) {
verbose("R%d partial copy of pointer\n", verbose(env,
"R%d partial copy of pointer\n",
insn->src_reg); insn->src_reg);
return -EACCES; return -EACCES;
} }
mark_reg_unknown(regs, insn->dst_reg); mark_reg_unknown(env, regs, insn->dst_reg);
/* high 32 bits are known zero. */ /* high 32 bits are known zero. */
regs[insn->dst_reg].var_off = tnum_cast( regs[insn->dst_reg].var_off = tnum_cast(
regs[insn->dst_reg].var_off, 4); regs[insn->dst_reg].var_off, 4);
...@@ -2430,14 +2456,14 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2430,14 +2456,14 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
} }
} else if (opcode > BPF_END) { } else if (opcode > BPF_END) {
verbose("invalid BPF_ALU opcode %x\n", opcode); verbose(env, "invalid BPF_ALU opcode %x\n", opcode);
return -EINVAL; return -EINVAL;
} else { /* all other ALU ops: and, sub, xor, add, ... */ } else { /* all other ALU ops: and, sub, xor, add, ... */
if (BPF_SRC(insn->code) == BPF_X) { if (BPF_SRC(insn->code) == BPF_X) {
if (insn->imm != 0 || insn->off != 0) { if (insn->imm != 0 || insn->off != 0) {
verbose("BPF_ALU uses reserved fields\n"); verbose(env, "BPF_ALU uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
/* check src1 operand */ /* check src1 operand */
...@@ -2446,7 +2472,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2446,7 +2472,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
return err; return err;
} else { } else {
if (insn->src_reg != BPF_REG_0 || insn->off != 0) { if (insn->src_reg != BPF_REG_0 || insn->off != 0) {
verbose("BPF_ALU uses reserved fields\n"); verbose(env, "BPF_ALU uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
} }
...@@ -2458,7 +2484,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2458,7 +2484,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
if ((opcode == BPF_MOD || opcode == BPF_DIV) && if ((opcode == BPF_MOD || opcode == BPF_DIV) &&
BPF_SRC(insn->code) == BPF_K && insn->imm == 0) { BPF_SRC(insn->code) == BPF_K && insn->imm == 0) {
verbose("div by zero\n"); verbose(env, "div by zero\n");
return -EINVAL; return -EINVAL;
} }
...@@ -2467,7 +2493,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2467,7 +2493,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
int size = BPF_CLASS(insn->code) == BPF_ALU64 ? 64 : 32; int size = BPF_CLASS(insn->code) == BPF_ALU64 ? 64 : 32;
if (insn->imm < 0 || insn->imm >= size) { if (insn->imm < 0 || insn->imm >= size) {
verbose("invalid shift %d\n", insn->imm); verbose(env, "invalid shift %d\n", insn->imm);
return -EINVAL; return -EINVAL;
} }
} }
...@@ -2820,13 +2846,13 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env, ...@@ -2820,13 +2846,13 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env,
int err; int err;
if (opcode > BPF_JSLE) { if (opcode > BPF_JSLE) {
verbose("invalid BPF_JMP opcode %x\n", opcode); verbose(env, "invalid BPF_JMP opcode %x\n", opcode);
return -EINVAL; return -EINVAL;
} }
if (BPF_SRC(insn->code) == BPF_X) { if (BPF_SRC(insn->code) == BPF_X) {
if (insn->imm != 0) { if (insn->imm != 0) {
verbose("BPF_JMP uses reserved fields\n"); verbose(env, "BPF_JMP uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -2836,13 +2862,13 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env, ...@@ -2836,13 +2862,13 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env,
return err; return err;
if (is_pointer_value(env, insn->src_reg)) { if (is_pointer_value(env, insn->src_reg)) {
verbose("R%d pointer comparison prohibited\n", verbose(env, "R%d pointer comparison prohibited\n",
insn->src_reg); insn->src_reg);
return -EACCES; return -EACCES;
} }
} else { } else {
if (insn->src_reg != BPF_REG_0) { if (insn->src_reg != BPF_REG_0) {
verbose("BPF_JMP uses reserved fields\n"); verbose(env, "BPF_JMP uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
} }
...@@ -2954,11 +2980,12 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env, ...@@ -2954,11 +2980,12 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env,
find_good_pkt_pointers(this_branch, &regs[insn->src_reg], find_good_pkt_pointers(this_branch, &regs[insn->src_reg],
PTR_TO_PACKET_META); PTR_TO_PACKET_META);
} else if (is_pointer_value(env, insn->dst_reg)) { } else if (is_pointer_value(env, insn->dst_reg)) {
verbose("R%d pointer comparison prohibited\n", insn->dst_reg); verbose(env, "R%d pointer comparison prohibited\n",
insn->dst_reg);
return -EACCES; return -EACCES;
} }
if (verifier_log.level) if (env->log.level)
print_verifier_state(this_branch); print_verifier_state(env, this_branch);
return 0; return 0;
} }
...@@ -2977,11 +3004,11 @@ static int check_ld_imm(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -2977,11 +3004,11 @@ static int check_ld_imm(struct bpf_verifier_env *env, struct bpf_insn *insn)
int err; int err;
if (BPF_SIZE(insn->code) != BPF_DW) { if (BPF_SIZE(insn->code) != BPF_DW) {
verbose("invalid BPF_LD_IMM insn\n"); verbose(env, "invalid BPF_LD_IMM insn\n");
return -EINVAL; return -EINVAL;
} }
if (insn->off != 0) { if (insn->off != 0) {
verbose("BPF_LD_IMM64 uses reserved fields\n"); verbose(env, "BPF_LD_IMM64 uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3039,14 +3066,14 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -3039,14 +3066,14 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn)
int i, err; int i, err;
if (!may_access_skb(env->prog->type)) { if (!may_access_skb(env->prog->type)) {
verbose("BPF_LD_[ABS|IND] instructions not allowed for this program type\n"); verbose(env, "BPF_LD_[ABS|IND] instructions not allowed for this program type\n");
return -EINVAL; return -EINVAL;
} }
if (insn->dst_reg != BPF_REG_0 || insn->off != 0 || if (insn->dst_reg != BPF_REG_0 || insn->off != 0 ||
BPF_SIZE(insn->code) == BPF_DW || BPF_SIZE(insn->code) == BPF_DW ||
(mode == BPF_ABS && insn->src_reg != BPF_REG_0)) { (mode == BPF_ABS && insn->src_reg != BPF_REG_0)) {
verbose("BPF_LD_[ABS|IND] uses reserved fields\n"); verbose(env, "BPF_LD_[ABS|IND] uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3056,7 +3083,8 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -3056,7 +3083,8 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn)
return err; return err;
if (regs[BPF_REG_6].type != PTR_TO_CTX) { if (regs[BPF_REG_6].type != PTR_TO_CTX) {
verbose("at the time of BPF_LD_ABS|IND R6 != pointer to skb\n"); verbose(env,
"at the time of BPF_LD_ABS|IND R6 != pointer to skb\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3069,7 +3097,7 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -3069,7 +3097,7 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn)
/* reset caller saved regs to unreadable */ /* reset caller saved regs to unreadable */
for (i = 0; i < CALLER_SAVED_REGS; i++) { for (i = 0; i < CALLER_SAVED_REGS; i++) {
mark_reg_not_init(regs, caller_saved[i]); mark_reg_not_init(env, regs, caller_saved[i]);
check_reg_arg(env, caller_saved[i], DST_OP_NO_MARK); check_reg_arg(env, caller_saved[i], DST_OP_NO_MARK);
} }
...@@ -3077,7 +3105,7 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn) ...@@ -3077,7 +3105,7 @@ static int check_ld_abs(struct bpf_verifier_env *env, struct bpf_insn *insn)
* the value fetched from the packet. * the value fetched from the packet.
* Already marked as written above. * Already marked as written above.
*/ */
mark_reg_unknown(regs, BPF_REG_0); mark_reg_unknown(env, regs, BPF_REG_0);
return 0; return 0;
} }
...@@ -3097,22 +3125,22 @@ static int check_return_code(struct bpf_verifier_env *env) ...@@ -3097,22 +3125,22 @@ static int check_return_code(struct bpf_verifier_env *env)
reg = &env->cur_state.regs[BPF_REG_0]; reg = &env->cur_state.regs[BPF_REG_0];
if (reg->type != SCALAR_VALUE) { if (reg->type != SCALAR_VALUE) {
verbose("At program exit the register R0 is not a known value (%s)\n", verbose(env, "At program exit the register R0 is not a known value (%s)\n",
reg_type_str[reg->type]); reg_type_str[reg->type]);
return -EINVAL; return -EINVAL;
} }
if (!tnum_in(range, reg->var_off)) { if (!tnum_in(range, reg->var_off)) {
verbose("At program exit the register R0 "); verbose(env, "At program exit the register R0 ");
if (!tnum_is_unknown(reg->var_off)) { if (!tnum_is_unknown(reg->var_off)) {
char tn_buf[48]; char tn_buf[48];
tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off); tnum_strn(tn_buf, sizeof(tn_buf), reg->var_off);
verbose("has value %s", tn_buf); verbose(env, "has value %s", tn_buf);
} else { } else {
verbose("has unknown scalar value"); verbose(env, "has unknown scalar value");
} }
verbose(" should have been 0 or 1\n"); verbose(env, " should have been 0 or 1\n");
return -EINVAL; return -EINVAL;
} }
return 0; return 0;
...@@ -3178,7 +3206,7 @@ static int push_insn(int t, int w, int e, struct bpf_verifier_env *env) ...@@ -3178,7 +3206,7 @@ static int push_insn(int t, int w, int e, struct bpf_verifier_env *env)
return 0; return 0;
if (w < 0 || w >= env->prog->len) { if (w < 0 || w >= env->prog->len) {
verbose("jump out of range from insn %d to %d\n", t, w); verbose(env, "jump out of range from insn %d to %d\n", t, w);
return -EINVAL; return -EINVAL;
} }
...@@ -3195,13 +3223,13 @@ static int push_insn(int t, int w, int e, struct bpf_verifier_env *env) ...@@ -3195,13 +3223,13 @@ static int push_insn(int t, int w, int e, struct bpf_verifier_env *env)
insn_stack[cur_stack++] = w; insn_stack[cur_stack++] = w;
return 1; return 1;
} else if ((insn_state[w] & 0xF0) == DISCOVERED) { } else if ((insn_state[w] & 0xF0) == DISCOVERED) {
verbose("back-edge from insn %d to %d\n", t, w); verbose(env, "back-edge from insn %d to %d\n", t, w);
return -EINVAL; return -EINVAL;
} else if (insn_state[w] == EXPLORED) { } else if (insn_state[w] == EXPLORED) {
/* forward- or cross-edge */ /* forward- or cross-edge */
insn_state[t] = DISCOVERED | e; insn_state[t] = DISCOVERED | e;
} else { } else {
verbose("insn state internal bug\n"); verbose(env, "insn state internal bug\n");
return -EFAULT; return -EFAULT;
} }
return 0; return 0;
...@@ -3295,7 +3323,7 @@ static int check_cfg(struct bpf_verifier_env *env) ...@@ -3295,7 +3323,7 @@ static int check_cfg(struct bpf_verifier_env *env)
mark_explored: mark_explored:
insn_state[t] = EXPLORED; insn_state[t] = EXPLORED;
if (cur_stack-- <= 0) { if (cur_stack-- <= 0) {
verbose("pop stack internal bug\n"); verbose(env, "pop stack internal bug\n");
ret = -EFAULT; ret = -EFAULT;
goto err_free; goto err_free;
} }
...@@ -3304,7 +3332,7 @@ static int check_cfg(struct bpf_verifier_env *env) ...@@ -3304,7 +3332,7 @@ static int check_cfg(struct bpf_verifier_env *env)
check_state: check_state:
for (i = 0; i < insn_cnt; i++) { for (i = 0; i < insn_cnt; i++) {
if (insn_state[i] != EXPLORED) { if (insn_state[i] != EXPLORED) {
verbose("unreachable insn %d\n", i); verbose(env, "unreachable insn %d\n", i);
ret = -EINVAL; ret = -EINVAL;
goto err_free; goto err_free;
} }
...@@ -3685,7 +3713,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3685,7 +3713,7 @@ static int do_check(struct bpf_verifier_env *env)
int insn_processed = 0; int insn_processed = 0;
bool do_print_state = false; bool do_print_state = false;
init_reg_state(regs); init_reg_state(env, regs);
state->parent = NULL; state->parent = NULL;
insn_idx = 0; insn_idx = 0;
for (;;) { for (;;) {
...@@ -3694,7 +3722,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3694,7 +3722,7 @@ static int do_check(struct bpf_verifier_env *env)
int err; int err;
if (insn_idx >= insn_cnt) { if (insn_idx >= insn_cnt) {
verbose("invalid insn idx %d insn_cnt %d\n", verbose(env, "invalid insn idx %d insn_cnt %d\n",
insn_idx, insn_cnt); insn_idx, insn_cnt);
return -EFAULT; return -EFAULT;
} }
...@@ -3703,7 +3731,8 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3703,7 +3731,8 @@ static int do_check(struct bpf_verifier_env *env)
class = BPF_CLASS(insn->code); class = BPF_CLASS(insn->code);
if (++insn_processed > BPF_COMPLEXITY_LIMIT_INSNS) { if (++insn_processed > BPF_COMPLEXITY_LIMIT_INSNS) {
verbose("BPF program is too large. Processed %d insn\n", verbose(env,
"BPF program is too large. Processed %d insn\n",
insn_processed); insn_processed);
return -E2BIG; return -E2BIG;
} }
...@@ -3713,12 +3742,12 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3713,12 +3742,12 @@ static int do_check(struct bpf_verifier_env *env)
return err; return err;
if (err == 1) { if (err == 1) {
/* found equivalent state, can prune the search */ /* found equivalent state, can prune the search */
if (verifier_log.level) { if (env->log.level) {
if (do_print_state) if (do_print_state)
verbose("\nfrom %d to %d: safe\n", verbose(env, "\nfrom %d to %d: safe\n",
prev_insn_idx, insn_idx); prev_insn_idx, insn_idx);
else else
verbose("%d: safe\n", insn_idx); verbose(env, "%d: safe\n", insn_idx);
} }
goto process_bpf_exit; goto process_bpf_exit;
} }
...@@ -3726,19 +3755,18 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3726,19 +3755,18 @@ static int do_check(struct bpf_verifier_env *env)
if (need_resched()) if (need_resched())
cond_resched(); cond_resched();
if (verifier_log.level > 1 || if (env->log.level > 1 || (env->log.level && do_print_state)) {
(verifier_log.level && do_print_state)) { if (env->log.level > 1)
if (verifier_log.level > 1) verbose(env, "%d:", insn_idx);
verbose("%d:", insn_idx);
else else
verbose("\nfrom %d to %d:", verbose(env, "\nfrom %d to %d:",
prev_insn_idx, insn_idx); prev_insn_idx, insn_idx);
print_verifier_state(&env->cur_state); print_verifier_state(env, &env->cur_state);
do_print_state = false; do_print_state = false;
} }
if (verifier_log.level) { if (env->log.level) {
verbose("%d: ", insn_idx); verbose(env, "%d: ", insn_idx);
print_bpf_insn(env, insn); print_bpf_insn(env, insn);
} }
...@@ -3795,7 +3823,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3795,7 +3823,7 @@ static int do_check(struct bpf_verifier_env *env)
* src_reg == stack|map in some other branch. * src_reg == stack|map in some other branch.
* Reject it. * Reject it.
*/ */
verbose("same insn cannot be used with different pointers\n"); verbose(env, "same insn cannot be used with different pointers\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3835,14 +3863,14 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3835,14 +3863,14 @@ static int do_check(struct bpf_verifier_env *env)
} else if (dst_reg_type != *prev_dst_type && } else if (dst_reg_type != *prev_dst_type &&
(dst_reg_type == PTR_TO_CTX || (dst_reg_type == PTR_TO_CTX ||
*prev_dst_type == PTR_TO_CTX)) { *prev_dst_type == PTR_TO_CTX)) {
verbose("same insn cannot be used with different pointers\n"); verbose(env, "same insn cannot be used with different pointers\n");
return -EINVAL; return -EINVAL;
} }
} else if (class == BPF_ST) { } else if (class == BPF_ST) {
if (BPF_MODE(insn->code) != BPF_MEM || if (BPF_MODE(insn->code) != BPF_MEM ||
insn->src_reg != BPF_REG_0) { insn->src_reg != BPF_REG_0) {
verbose("BPF_ST uses reserved fields\n"); verbose(env, "BPF_ST uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
/* check src operand */ /* check src operand */
...@@ -3865,7 +3893,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3865,7 +3893,7 @@ static int do_check(struct bpf_verifier_env *env)
insn->off != 0 || insn->off != 0 ||
insn->src_reg != BPF_REG_0 || insn->src_reg != BPF_REG_0 ||
insn->dst_reg != BPF_REG_0) { insn->dst_reg != BPF_REG_0) {
verbose("BPF_CALL uses reserved fields\n"); verbose(env, "BPF_CALL uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3878,7 +3906,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3878,7 +3906,7 @@ static int do_check(struct bpf_verifier_env *env)
insn->imm != 0 || insn->imm != 0 ||
insn->src_reg != BPF_REG_0 || insn->src_reg != BPF_REG_0 ||
insn->dst_reg != BPF_REG_0) { insn->dst_reg != BPF_REG_0) {
verbose("BPF_JA uses reserved fields\n"); verbose(env, "BPF_JA uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3890,7 +3918,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3890,7 +3918,7 @@ static int do_check(struct bpf_verifier_env *env)
insn->imm != 0 || insn->imm != 0 ||
insn->src_reg != BPF_REG_0 || insn->src_reg != BPF_REG_0 ||
insn->dst_reg != BPF_REG_0) { insn->dst_reg != BPF_REG_0) {
verbose("BPF_EXIT uses reserved fields\n"); verbose(env, "BPF_EXIT uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -3905,7 +3933,7 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3905,7 +3933,7 @@ static int do_check(struct bpf_verifier_env *env)
return err; return err;
if (is_pointer_value(env, BPF_REG_0)) { if (is_pointer_value(env, BPF_REG_0)) {
verbose("R0 leaks addr as return value\n"); verbose(env, "R0 leaks addr as return value\n");
return -EACCES; return -EACCES;
} }
...@@ -3940,19 +3968,19 @@ static int do_check(struct bpf_verifier_env *env) ...@@ -3940,19 +3968,19 @@ static int do_check(struct bpf_verifier_env *env)
insn_idx++; insn_idx++;
} else { } else {
verbose("invalid BPF_LD mode\n"); verbose(env, "invalid BPF_LD mode\n");
return -EINVAL; return -EINVAL;
} }
} else { } else {
verbose("unknown insn class %d\n", class); verbose(env, "unknown insn class %d\n", class);
return -EINVAL; return -EINVAL;
} }
insn_idx++; insn_idx++;
} }
verbose("processed %d insns, stack depth %d\n", verbose(env, "processed %d insns, stack depth %d\n", insn_processed,
insn_processed, env->prog->aux->stack_depth); env->prog->aux->stack_depth);
return 0; return 0;
} }
...@@ -3964,7 +3992,8 @@ static int check_map_prealloc(struct bpf_map *map) ...@@ -3964,7 +3992,8 @@ static int check_map_prealloc(struct bpf_map *map)
!(map->map_flags & BPF_F_NO_PREALLOC); !(map->map_flags & BPF_F_NO_PREALLOC);
} }
static int check_map_prog_compatibility(struct bpf_map *map, static int check_map_prog_compatibility(struct bpf_verifier_env *env,
struct bpf_map *map,
struct bpf_prog *prog) struct bpf_prog *prog)
{ {
...@@ -3975,12 +4004,12 @@ static int check_map_prog_compatibility(struct bpf_map *map, ...@@ -3975,12 +4004,12 @@ static int check_map_prog_compatibility(struct bpf_map *map,
*/ */
if (prog->type == BPF_PROG_TYPE_PERF_EVENT) { if (prog->type == BPF_PROG_TYPE_PERF_EVENT) {
if (!check_map_prealloc(map)) { if (!check_map_prealloc(map)) {
verbose("perf_event programs can only use preallocated hash map\n"); verbose(env, "perf_event programs can only use preallocated hash map\n");
return -EINVAL; return -EINVAL;
} }
if (map->inner_map_meta && if (map->inner_map_meta &&
!check_map_prealloc(map->inner_map_meta)) { !check_map_prealloc(map->inner_map_meta)) {
verbose("perf_event programs can only use preallocated inner hash map\n"); verbose(env, "perf_event programs can only use preallocated inner hash map\n");
return -EINVAL; return -EINVAL;
} }
} }
...@@ -4003,14 +4032,14 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env) ...@@ -4003,14 +4032,14 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env)
for (i = 0; i < insn_cnt; i++, insn++) { for (i = 0; i < insn_cnt; i++, insn++) {
if (BPF_CLASS(insn->code) == BPF_LDX && if (BPF_CLASS(insn->code) == BPF_LDX &&
(BPF_MODE(insn->code) != BPF_MEM || insn->imm != 0)) { (BPF_MODE(insn->code) != BPF_MEM || insn->imm != 0)) {
verbose("BPF_LDX uses reserved fields\n"); verbose(env, "BPF_LDX uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
if (BPF_CLASS(insn->code) == BPF_STX && if (BPF_CLASS(insn->code) == BPF_STX &&
((BPF_MODE(insn->code) != BPF_MEM && ((BPF_MODE(insn->code) != BPF_MEM &&
BPF_MODE(insn->code) != BPF_XADD) || insn->imm != 0)) { BPF_MODE(insn->code) != BPF_XADD) || insn->imm != 0)) {
verbose("BPF_STX uses reserved fields\n"); verbose(env, "BPF_STX uses reserved fields\n");
return -EINVAL; return -EINVAL;
} }
...@@ -4021,7 +4050,7 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env) ...@@ -4021,7 +4050,7 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env)
if (i == insn_cnt - 1 || insn[1].code != 0 || if (i == insn_cnt - 1 || insn[1].code != 0 ||
insn[1].dst_reg != 0 || insn[1].src_reg != 0 || insn[1].dst_reg != 0 || insn[1].src_reg != 0 ||
insn[1].off != 0) { insn[1].off != 0) {
verbose("invalid bpf_ld_imm64 insn\n"); verbose(env, "invalid bpf_ld_imm64 insn\n");
return -EINVAL; return -EINVAL;
} }
...@@ -4030,19 +4059,20 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env) ...@@ -4030,19 +4059,20 @@ static int replace_map_fd_with_map_ptr(struct bpf_verifier_env *env)
goto next_insn; goto next_insn;
if (insn->src_reg != BPF_PSEUDO_MAP_FD) { if (insn->src_reg != BPF_PSEUDO_MAP_FD) {
verbose("unrecognized bpf_ld_imm64 insn\n"); verbose(env,
"unrecognized bpf_ld_imm64 insn\n");
return -EINVAL; return -EINVAL;
} }
f = fdget(insn->imm); f = fdget(insn->imm);
map = __bpf_map_get(f); map = __bpf_map_get(f);
if (IS_ERR(map)) { if (IS_ERR(map)) {
verbose("fd %d is not pointing to valid bpf_map\n", verbose(env, "fd %d is not pointing to valid bpf_map\n",
insn->imm); insn->imm);
return PTR_ERR(map); return PTR_ERR(map);
} }
err = check_map_prog_compatibility(map, env->prog); err = check_map_prog_compatibility(env, map, env->prog);
if (err) { if (err) {
fdput(f); fdput(f);
return err; return err;
...@@ -4164,7 +4194,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env) ...@@ -4164,7 +4194,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env)
cnt = ops->gen_prologue(insn_buf, env->seen_direct_write, cnt = ops->gen_prologue(insn_buf, env->seen_direct_write,
env->prog); env->prog);
if (cnt >= ARRAY_SIZE(insn_buf)) { if (cnt >= ARRAY_SIZE(insn_buf)) {
verbose("bpf verifier is misconfigured\n"); verbose(env, "bpf verifier is misconfigured\n");
return -EINVAL; return -EINVAL;
} else if (cnt) { } else if (cnt) {
new_prog = bpf_patch_insn_data(env, 0, insn_buf, cnt); new_prog = bpf_patch_insn_data(env, 0, insn_buf, cnt);
...@@ -4212,7 +4242,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env) ...@@ -4212,7 +4242,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env)
u8 size_code; u8 size_code;
if (type == BPF_WRITE) { if (type == BPF_WRITE) {
verbose("bpf verifier narrow ctx access misconfigured\n"); verbose(env, "bpf verifier narrow ctx access misconfigured\n");
return -EINVAL; return -EINVAL;
} }
...@@ -4231,7 +4261,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env) ...@@ -4231,7 +4261,7 @@ static int convert_ctx_accesses(struct bpf_verifier_env *env)
&target_size); &target_size);
if (cnt == 0 || cnt >= ARRAY_SIZE(insn_buf) || if (cnt == 0 || cnt >= ARRAY_SIZE(insn_buf) ||
(ctx_field_size && !target_size)) { (ctx_field_size && !target_size)) {
verbose("bpf verifier is misconfigured\n"); verbose(env, "bpf verifier is misconfigured\n");
return -EINVAL; return -EINVAL;
} }
...@@ -4313,7 +4343,7 @@ static int fixup_bpf_calls(struct bpf_verifier_env *env) ...@@ -4313,7 +4343,7 @@ static int fixup_bpf_calls(struct bpf_verifier_env *env)
cnt = map_ptr->ops->map_gen_lookup(map_ptr, insn_buf); cnt = map_ptr->ops->map_gen_lookup(map_ptr, insn_buf);
if (cnt == 0 || cnt >= ARRAY_SIZE(insn_buf)) { if (cnt == 0 || cnt >= ARRAY_SIZE(insn_buf)) {
verbose("bpf verifier is misconfigured\n"); verbose(env, "bpf verifier is misconfigured\n");
return -EINVAL; return -EINVAL;
} }
...@@ -4357,7 +4387,8 @@ static int fixup_bpf_calls(struct bpf_verifier_env *env) ...@@ -4357,7 +4387,8 @@ static int fixup_bpf_calls(struct bpf_verifier_env *env)
* programs to call them, must be real in-kernel functions * programs to call them, must be real in-kernel functions
*/ */
if (!fn->func) { if (!fn->func) {
verbose("kernel subsystem misconfigured func %s#%d\n", verbose(env,
"kernel subsystem misconfigured func %s#%d\n",
func_id_name(insn->imm), insn->imm); func_id_name(insn->imm), insn->imm);
return -EFAULT; return -EFAULT;
} }
...@@ -4391,8 +4422,8 @@ static void free_states(struct bpf_verifier_env *env) ...@@ -4391,8 +4422,8 @@ static void free_states(struct bpf_verifier_env *env)
int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) int bpf_check(struct bpf_prog **prog, union bpf_attr *attr)
{ {
struct bpf_verifer_log *log = &verifier_log;
struct bpf_verifier_env *env; struct bpf_verifier_env *env;
struct bpf_verifer_log *log;
int ret = -EINVAL; int ret = -EINVAL;
/* 'struct bpf_verifier_env' can be global, but since it's not small, /* 'struct bpf_verifier_env' can be global, but since it's not small,
...@@ -4401,6 +4432,7 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) ...@@ -4401,6 +4432,7 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr)
env = kzalloc(sizeof(struct bpf_verifier_env), GFP_KERNEL); env = kzalloc(sizeof(struct bpf_verifier_env), GFP_KERNEL);
if (!env) if (!env)
return -ENOMEM; return -ENOMEM;
log = &env->log;
env->insn_aux_data = vzalloc(sizeof(struct bpf_insn_aux_data) * env->insn_aux_data = vzalloc(sizeof(struct bpf_insn_aux_data) *
(*prog)->len); (*prog)->len);
...@@ -4419,7 +4451,6 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) ...@@ -4419,7 +4451,6 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr)
log->level = attr->log_level; log->level = attr->log_level;
log->ubuf = (char __user *) (unsigned long) attr->log_buf; log->ubuf = (char __user *) (unsigned long) attr->log_buf;
log->len_total = attr->log_size; log->len_total = attr->log_size;
log->len_used = 0;
ret = -EINVAL; ret = -EINVAL;
/* log attributes have to be sane */ /* log attributes have to be sane */
...@@ -4431,8 +4462,6 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) ...@@ -4431,8 +4462,6 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr)
log->kbuf = vmalloc(log->len_total); log->kbuf = vmalloc(log->len_total);
if (!log->kbuf) if (!log->kbuf)
goto err_unlock; goto err_unlock;
} else {
log->level = 0;
} }
env->strict_alignment = !!(attr->prog_flags & BPF_F_STRICT_ALIGNMENT); env->strict_alignment = !!(attr->prog_flags & BPF_F_STRICT_ALIGNMENT);
...@@ -4543,8 +4572,6 @@ int bpf_analyzer(struct bpf_prog *prog, const struct bpf_ext_analyzer_ops *ops, ...@@ -4543,8 +4572,6 @@ int bpf_analyzer(struct bpf_prog *prog, const struct bpf_ext_analyzer_ops *ops,
/* grab the mutex to protect few globals used by verifier */ /* grab the mutex to protect few globals used by verifier */
mutex_lock(&bpf_verifier_lock); mutex_lock(&bpf_verifier_lock);
verifier_log.level = 0;
env->strict_alignment = false; env->strict_alignment = false;
if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS))
env->strict_alignment = true; env->strict_alignment = true;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment