Commit e42fc3c2 authored by Johan Almbladh's avatar Johan Almbladh Committed by Daniel Borkmann

bpf/tests: Minor restructuring of ALU tests

This patch moves the ALU LSH/RSH/ARSH reference computations into the
common reference value function. Also fix typo in constants so they
now have the intended values.
Signed-off-by: default avatarJohan Almbladh <johan.almbladh@anyfinetworks.com>
Signed-off-by: default avatarDaniel Borkmann <daniel@iogearbox.net>
Link: https://lore.kernel.org/bpf/20211001130348.3670534-7-johan.almbladh@anyfinetworks.com
parent e2f9797b
...@@ -538,6 +538,57 @@ static int bpf_fill_max_jmp_never_taken(struct bpf_test *self) ...@@ -538,6 +538,57 @@ static int bpf_fill_max_jmp_never_taken(struct bpf_test *self)
return __bpf_fill_max_jmp(self, BPF_JLT, 0); return __bpf_fill_max_jmp(self, BPF_JLT, 0);
} }
/* ALU result computation used in tests */
static bool __bpf_alu_result(u64 *res, u64 v1, u64 v2, u8 op)
{
*res = 0;
switch (op) {
case BPF_MOV:
*res = v2;
break;
case BPF_AND:
*res = v1 & v2;
break;
case BPF_OR:
*res = v1 | v2;
break;
case BPF_XOR:
*res = v1 ^ v2;
break;
case BPF_LSH:
*res = v1 << v2;
break;
case BPF_RSH:
*res = v1 >> v2;
break;
case BPF_ARSH:
*res = v1 >> v2;
if (v2 > 0 && v1 > S64_MAX)
*res |= ~0ULL << (64 - v2);
break;
case BPF_ADD:
*res = v1 + v2;
break;
case BPF_SUB:
*res = v1 - v2;
break;
case BPF_MUL:
*res = v1 * v2;
break;
case BPF_DIV:
if (v2 == 0)
return false;
*res = div64_u64(v1, v2);
break;
case BPF_MOD:
if (v2 == 0)
return false;
div64_u64_rem(v1, v2, res);
break;
}
return true;
}
/* Test an ALU shift operation for all valid shift values */ /* Test an ALU shift operation for all valid shift values */
static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op, static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op,
u8 mode, bool alu32) u8 mode, bool alu32)
...@@ -576,37 +627,19 @@ static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op, ...@@ -576,37 +627,19 @@ static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op,
insn[i++] = BPF_ALU32_IMM(op, R1, imm); insn[i++] = BPF_ALU32_IMM(op, R1, imm);
else else
insn[i++] = BPF_ALU32_REG(op, R1, R2); insn[i++] = BPF_ALU32_REG(op, R1, R2);
switch (op) {
case BPF_LSH: if (op == BPF_ARSH)
val = (u32)reg << imm; reg = (s32)reg;
break; else
case BPF_RSH: reg = (u32)reg;
val = (u32)reg >> imm; __bpf_alu_result(&val, reg, imm, op);
break; val = (u32)val;
case BPF_ARSH:
val = (u32)reg >> imm;
if (imm > 0 && (reg & 0x80000000))
val |= ~(u32)0 << (32 - imm);
break;
}
} else { } else {
if (mode == BPF_K) if (mode == BPF_K)
insn[i++] = BPF_ALU64_IMM(op, R1, imm); insn[i++] = BPF_ALU64_IMM(op, R1, imm);
else else
insn[i++] = BPF_ALU64_REG(op, R1, R2); insn[i++] = BPF_ALU64_REG(op, R1, R2);
switch (op) { __bpf_alu_result(&val, reg, imm, op);
case BPF_LSH:
val = (u64)reg << imm;
break;
case BPF_RSH:
val = (u64)reg >> imm;
break;
case BPF_ARSH:
val = (u64)reg >> imm;
if (imm > 0 && reg < 0)
val |= ~(u64)0 << (64 - imm);
break;
}
} }
/* /*
...@@ -799,46 +832,6 @@ static int __bpf_fill_pattern(struct bpf_test *self, void *arg, ...@@ -799,46 +832,6 @@ static int __bpf_fill_pattern(struct bpf_test *self, void *arg,
* test is designed to verify e.g. the ALU and ALU64 operations for JITs that * test is designed to verify e.g. the ALU and ALU64 operations for JITs that
* emit different code depending on the magnitude of the immediate value. * emit different code depending on the magnitude of the immediate value.
*/ */
static bool __bpf_alu_result(u64 *res, u64 v1, u64 v2, u8 op)
{
*res = 0;
switch (op) {
case BPF_MOV:
*res = v2;
break;
case BPF_AND:
*res = v1 & v2;
break;
case BPF_OR:
*res = v1 | v2;
break;
case BPF_XOR:
*res = v1 ^ v2;
break;
case BPF_ADD:
*res = v1 + v2;
break;
case BPF_SUB:
*res = v1 - v2;
break;
case BPF_MUL:
*res = v1 * v2;
break;
case BPF_DIV:
if (v2 == 0)
return false;
*res = div64_u64(v1, v2);
break;
case BPF_MOD:
if (v2 == 0)
return false;
div64_u64_rem(v1, v2, res);
break;
}
return true;
}
static int __bpf_emit_alu64_imm(struct bpf_test *self, void *arg, static int __bpf_emit_alu64_imm(struct bpf_test *self, void *arg,
struct bpf_insn *insns, s64 dst, s64 imm) struct bpf_insn *insns, s64 dst, s64 imm)
{ {
...@@ -7881,7 +7874,7 @@ static struct bpf_test tests[] = { ...@@ -7881,7 +7874,7 @@ static struct bpf_test tests[] = {
"BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful return", "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful return",
.u.insns_int = { .u.insns_int = {
BPF_LD_IMM64(R1, 0x0123456789abcdefULL), BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
BPF_LD_IMM64(R2, 0xfecdba9876543210ULL), BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
BPF_ALU64_REG(BPF_MOV, R0, R1), BPF_ALU64_REG(BPF_MOV, R0, R1),
BPF_STX_MEM(BPF_DW, R10, R1, -40), BPF_STX_MEM(BPF_DW, R10, R1, -40),
BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
...@@ -7898,7 +7891,7 @@ static struct bpf_test tests[] = { ...@@ -7898,7 +7891,7 @@ static struct bpf_test tests[] = {
"BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful store", "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test successful store",
.u.insns_int = { .u.insns_int = {
BPF_LD_IMM64(R1, 0x0123456789abcdefULL), BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
BPF_LD_IMM64(R2, 0xfecdba9876543210ULL), BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
BPF_ALU64_REG(BPF_MOV, R0, R1), BPF_ALU64_REG(BPF_MOV, R0, R1),
BPF_STX_MEM(BPF_DW, R10, R0, -40), BPF_STX_MEM(BPF_DW, R10, R0, -40),
BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
...@@ -7916,7 +7909,7 @@ static struct bpf_test tests[] = { ...@@ -7916,7 +7909,7 @@ static struct bpf_test tests[] = {
"BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure return", "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure return",
.u.insns_int = { .u.insns_int = {
BPF_LD_IMM64(R1, 0x0123456789abcdefULL), BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
BPF_LD_IMM64(R2, 0xfecdba9876543210ULL), BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
BPF_ALU64_REG(BPF_MOV, R0, R1), BPF_ALU64_REG(BPF_MOV, R0, R1),
BPF_ALU64_IMM(BPF_ADD, R0, 1), BPF_ALU64_IMM(BPF_ADD, R0, 1),
BPF_STX_MEM(BPF_DW, R10, R1, -40), BPF_STX_MEM(BPF_DW, R10, R1, -40),
...@@ -7934,7 +7927,7 @@ static struct bpf_test tests[] = { ...@@ -7934,7 +7927,7 @@ static struct bpf_test tests[] = {
"BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure store", "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test failure store",
.u.insns_int = { .u.insns_int = {
BPF_LD_IMM64(R1, 0x0123456789abcdefULL), BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
BPF_LD_IMM64(R2, 0xfecdba9876543210ULL), BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
BPF_ALU64_REG(BPF_MOV, R0, R1), BPF_ALU64_REG(BPF_MOV, R0, R1),
BPF_ALU64_IMM(BPF_ADD, R0, 1), BPF_ALU64_IMM(BPF_ADD, R0, 1),
BPF_STX_MEM(BPF_DW, R10, R1, -40), BPF_STX_MEM(BPF_DW, R10, R1, -40),
...@@ -7953,11 +7946,11 @@ static struct bpf_test tests[] = { ...@@ -7953,11 +7946,11 @@ static struct bpf_test tests[] = {
"BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test side effects", "BPF_ATOMIC | BPF_DW, BPF_CMPXCHG: Test side effects",
.u.insns_int = { .u.insns_int = {
BPF_LD_IMM64(R1, 0x0123456789abcdefULL), BPF_LD_IMM64(R1, 0x0123456789abcdefULL),
BPF_LD_IMM64(R2, 0xfecdba9876543210ULL), BPF_LD_IMM64(R2, 0xfedcba9876543210ULL),
BPF_ALU64_REG(BPF_MOV, R0, R1), BPF_ALU64_REG(BPF_MOV, R0, R1),
BPF_STX_MEM(BPF_DW, R10, R1, -40), BPF_STX_MEM(BPF_DW, R10, R1, -40),
BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40), BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, R10, R2, -40),
BPF_LD_IMM64(R0, 0xfecdba9876543210ULL), BPF_LD_IMM64(R0, 0xfedcba9876543210ULL),
BPF_JMP_REG(BPF_JNE, R0, R2, 1), BPF_JMP_REG(BPF_JNE, R0, R2, 1),
BPF_ALU64_REG(BPF_SUB, R0, R2), BPF_ALU64_REG(BPF_SUB, R0, R2),
BPF_EXIT_INSN(), BPF_EXIT_INSN(),
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment