Lines Matching refs:DST

54 #define DST	regs[insn->dst_reg]  macro
1424 DST = DST OP (SRC & 63); \ in ___bpf_prog_run()
1427 DST = (u32) DST OP ((u32) SRC & 31); \ in ___bpf_prog_run()
1430 DST = DST OP IMM; \ in ___bpf_prog_run()
1433 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1438 DST = DST OP SRC; \ in ___bpf_prog_run()
1441 DST = (u32) DST OP (u32) SRC; \ in ___bpf_prog_run()
1444 DST = DST OP IMM; \ in ___bpf_prog_run()
1447 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1460 DST = (u32) -DST; in ___bpf_prog_run()
1463 DST = -DST; in ___bpf_prog_run()
1466 DST = (u32) SRC; in ___bpf_prog_run()
1469 DST = (u32) IMM; in ___bpf_prog_run()
1472 DST = SRC; in ___bpf_prog_run()
1475 DST = IMM; in ___bpf_prog_run()
1478 DST = (u64) (u32) insn[0].imm | ((u64) (u32) insn[1].imm) << 32; in ___bpf_prog_run()
1482 DST = (u64) (u32) (((s32) DST) >> (SRC & 31)); in ___bpf_prog_run()
1485 DST = (u64) (u32) (((s32) DST) >> IMM); in ___bpf_prog_run()
1488 (*(s64 *) &DST) >>= (SRC & 63); in ___bpf_prog_run()
1491 (*(s64 *) &DST) >>= IMM; in ___bpf_prog_run()
1494 div64_u64_rem(DST, SRC, &AX); in ___bpf_prog_run()
1495 DST = AX; in ___bpf_prog_run()
1498 AX = (u32) DST; in ___bpf_prog_run()
1499 DST = do_div(AX, (u32) SRC); in ___bpf_prog_run()
1502 div64_u64_rem(DST, IMM, &AX); in ___bpf_prog_run()
1503 DST = AX; in ___bpf_prog_run()
1506 AX = (u32) DST; in ___bpf_prog_run()
1507 DST = do_div(AX, (u32) IMM); in ___bpf_prog_run()
1510 DST = div64_u64(DST, SRC); in ___bpf_prog_run()
1513 AX = (u32) DST; in ___bpf_prog_run()
1515 DST = (u32) AX; in ___bpf_prog_run()
1518 DST = div64_u64(DST, IMM); in ___bpf_prog_run()
1521 AX = (u32) DST; in ___bpf_prog_run()
1523 DST = (u32) AX; in ___bpf_prog_run()
1528 DST = (__force u16) cpu_to_be16(DST); in ___bpf_prog_run()
1531 DST = (__force u32) cpu_to_be32(DST); in ___bpf_prog_run()
1534 DST = (__force u64) cpu_to_be64(DST); in ___bpf_prog_run()
1541 DST = (__force u16) cpu_to_le16(DST); in ___bpf_prog_run()
1544 DST = (__force u32) cpu_to_le32(DST); in ___bpf_prog_run()
1547 DST = (__force u64) cpu_to_le64(DST); in ___bpf_prog_run()
1604 if ((SIGN##64) DST CMP_OP (SIGN##64) SRC) { \ in ___bpf_prog_run()
1610 if ((SIGN##32) DST CMP_OP (SIGN##32) SRC) { \ in ___bpf_prog_run()
1616 if ((SIGN##64) DST CMP_OP (SIGN##64) IMM) { \ in ___bpf_prog_run()
1622 if ((SIGN##32) DST CMP_OP (SIGN##32) IMM) { \ in ___bpf_prog_run()
1656 *(SIZE *)(unsigned long) (DST + insn->off) = SRC; \ in ___bpf_prog_run()
1659 *(SIZE *)(unsigned long) (DST + insn->off) = IMM; \ in ___bpf_prog_run()
1662 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
1672 bpf_probe_read_kernel(&DST, SIZE, (const void *)(long) (SRC + insn->off)); \ in ___bpf_prog_run()
1684 (DST + insn->off)); \ in ___bpf_prog_run()
1687 (DST + insn->off)); \ in ___bpf_prog_run()
1693 (atomic_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
1697 (atomic64_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
1712 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1716 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1722 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1726 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()