Coverage Report

Created: 2026-03-13 06:50

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/src/capstonenext/arch/BPF/BPFDisassembler.c
Line
Count
Source
1
/* Capstone Disassembly Engine */
2
/* BPF Backend by david942j <david942j@gmail.com>, 2019 */
3
/* SPDX-FileCopyrightText: 2024 Roee Toledano <roeetoledano10@gmail.com> */
4
/* SPDX-License-Identifier: BSD-3 */
5
6
#ifdef CAPSTONE_HAS_BPF
7
8
#include <string.h>
9
#include <stddef.h> // offsetof macro
10
11
#include "BPFConstants.h"
12
#include "BPFDisassembler.h"
13
#include "BPFMapping.h"
14
#include "../../Mapping.h"
15
#include "../../cs_priv.h"
16
#include "../../utils.h"
17
18
///< Malloc bpf_internal, also checks if code_len is large enough.
19
static bpf_internal *alloc_bpf_internal(const size_t code_len)
20
17.2k
{
21
17.2k
  bpf_internal *bpf;
22
23
17.2k
  if (code_len < 8)
24
265
    return NULL;
25
16.9k
  bpf = cs_mem_malloc(sizeof(bpf_internal));
26
16.9k
  if (bpf == NULL)
27
0
    return NULL;
28
  /* default value */
29
16.9k
  bpf->insn_size = 8;
30
16.9k
  return bpf;
31
16.9k
}
32
33
///< Fetch a cBPF structure from code
34
static bpf_internal *fetch_cbpf(MCInst *instr, const uint8_t *code,
35
        const size_t code_len)
36
6.66k
{
37
6.66k
  bpf_internal *bpf;
38
39
6.66k
  bpf = alloc_bpf_internal(code_len);
40
6.66k
  if (bpf == NULL)
41
81
    return NULL;
42
43
6.58k
  bpf->op = readBytes16(instr, code);
44
6.58k
  bpf->jt = code[2];
45
6.58k
  bpf->jf = code[3];
46
6.58k
  bpf->k = readBytes32(instr, code + 4);
47
6.58k
  return bpf;
48
6.66k
}
49
50
///< Fetch an eBPF structure from code
51
static bpf_internal *fetch_ebpf(MCInst *instr, const uint8_t *code,
52
        const size_t code_len)
53
10.5k
{
54
10.5k
  bpf_internal *bpf;
55
56
10.5k
  bpf = alloc_bpf_internal(code_len);
57
10.5k
  if (bpf == NULL)
58
184
    return NULL;
59
60
10.3k
  bpf->op = (uint16_t)code[0];
61
10.3k
  bpf->dst = code[1] & 0xf;
62
10.3k
  bpf->src = (code[1] & 0xf0) >> 4;
63
64
  // eBPF has one 16-byte instruction: BPF_LD | BPF_DW | BPF_IMM,
65
  // in this case imm is combined with the next block's imm.
66
10.3k
  if (bpf->op == (BPF_CLASS_LD | BPF_SIZE_DW | BPF_MODE_IMM)) {
67
92
    if (code_len < 16) {
68
2
      cs_mem_free(bpf);
69
2
      return NULL;
70
2
    }
71
90
    bpf->k = readBytes32(instr, code + 4) |
72
90
       (((uint64_t)readBytes32(instr, code + 12)) << 32);
73
90
    bpf->insn_size = 16;
74
10.2k
  } else {
75
10.2k
    bpf->offset = readBytes16(instr, code + 2);
76
10.2k
    bpf->k = readBytes32(instr, code + 4);
77
10.2k
  }
78
10.3k
  return bpf;
79
10.3k
}
80
81
#define CHECK_READABLE_REG(ud, reg) \
82
6.70k
  do { \
83
6.70k
    if (!((reg) >= BPF_REG_R0 && (reg) <= BPF_REG_R10)) \
84
6.70k
      return false; \
85
6.70k
  } while (0)
86
87
#define CHECK_WRITEABLE_REG(ud, reg) \
88
3.60k
  do { \
89
3.60k
    if (!((reg) >= BPF_REG_R0 && (reg) < BPF_REG_R10)) \
90
3.60k
      return false; \
91
3.60k
  } while (0)
92
93
#define CHECK_READABLE_AND_PUSH(ud, MI, r) \
94
6.70k
  do { \
95
6.70k
    CHECK_READABLE_REG(ud, r + BPF_REG_R0); \
96
6.70k
    MCOperand_CreateReg0(MI, r + BPF_REG_R0); \
97
6.67k
  } while (0)
98
99
#define CHECK_WRITABLE_AND_PUSH(ud, MI, r) \
100
3.60k
  do { \
101
3.60k
    CHECK_WRITEABLE_REG(ud, r + BPF_REG_R0); \
102
3.60k
    MCOperand_CreateReg0(MI, r + BPF_REG_R0); \
103
3.59k
  } while (0)
104
105
static bool decodeLoad(MCInst *MI, bpf_internal *bpf)
106
5.07k
{
107
5.07k
  if (!EBPF_MODE(MI->csh->mode)) {
108
    /*
109
     *  +-----+-----------+--------------------+
110
     *  | ldb |    [k]    |       [x+k]        |
111
     *  | ldh |    [k]    |       [x+k]        |
112
     *  +-----+-----------+--------------------+
113
     */
114
2.98k
    if (BPF_SIZE(bpf->op) == BPF_SIZE_DW)
115
4
      return false;
116
2.98k
    if (BPF_SIZE(bpf->op) == BPF_SIZE_B ||
117
2.14k
        BPF_SIZE(bpf->op) == BPF_SIZE_H) {
118
      /* no ldx */
119
933
      if (BPF_CLASS(bpf->op) != BPF_CLASS_LD)
120
6
        return false;
121
      /* can only be BPF_ABS and BPF_IND */
122
927
      if (BPF_MODE(bpf->op) == BPF_MODE_ABS) {
123
252
        MCOperand_CreateImm0(MI, bpf->k);
124
252
        return true;
125
675
      } else if (BPF_MODE(bpf->op) == BPF_MODE_IND) {
126
673
        MCOperand_CreateReg0(MI, BPF_REG_X);
127
673
        MCOperand_CreateImm0(MI, bpf->k);
128
673
        return true;
129
673
      }
130
2
      return false;
131
927
    }
132
    /*
133
     *  +-----+----+------+------+-----+-------+
134
     *  | ld  | #k | #len | M[k] | [k] | [x+k] |
135
     *  +-----+----+------+------+-----+-------+
136
     *  | ldx | #k | #len | M[k] | 4*([k]&0xf) |
137
     *  +-----+----+------+------+-------------+
138
     */
139
2.05k
    switch (BPF_MODE(bpf->op)) {
140
1.26k
    default:
141
1.26k
      break;
142
1.26k
    case BPF_MODE_IMM:
143
634
      MCOperand_CreateImm0(MI, bpf->k);
144
634
      return true;
145
99
    case BPF_MODE_LEN:
146
99
      return true;
147
49
    case BPF_MODE_MEM:
148
49
      MCOperand_CreateImm0(MI, bpf->k);
149
49
      return true;
150
2.05k
    }
151
1.26k
    if (BPF_CLASS(bpf->op) == BPF_CLASS_LD) {
152
1.07k
      if (BPF_MODE(bpf->op) == BPF_MODE_ABS) {
153
406
        MCOperand_CreateImm0(MI, bpf->k);
154
406
        return true;
155
667
      } else if (BPF_MODE(bpf->op) == BPF_MODE_IND) {
156
665
        MCOperand_CreateReg0(MI, BPF_REG_X);
157
665
        MCOperand_CreateImm0(MI, bpf->k);
158
665
        return true;
159
665
      }
160
1.07k
    } else { /* LDX */
161
195
      if (BPF_MODE(bpf->op) == BPF_MODE_MSH) {
162
194
        MCOperand_CreateImm0(MI, bpf->k);
163
194
        return true;
164
194
      }
165
195
    }
166
3
    return false;
167
1.26k
  }
168
169
  /* eBPF mode */
170
  /*
171
   * - IMM: lddw dst, imm64
172
   * - ABS: ld{w,h,b} [k]
173
   * - IND: ld{w,h,b} [src]
174
   * - MEM: ldx{w,h,b,dw} dst, [src+off]
175
   */
176
2.08k
  if (BPF_CLASS(bpf->op) == BPF_CLASS_LD) {
177
1.26k
    switch (BPF_MODE(bpf->op)) {
178
129
    case BPF_MODE_IMM:
179
129
      if (bpf->op !=
180
129
          (BPF_CLASS_LD | BPF_SIZE_DW | BPF_MODE_IMM))
181
39
        return false;
182
90
      CHECK_WRITABLE_AND_PUSH(ud, MI, bpf->dst);
183
90
      MCOperand_CreateImm0(MI, bpf->k);
184
90
      return true;
185
800
    case BPF_MODE_ABS:
186
800
      MCOperand_CreateImm0(MI, bpf->k);
187
800
      return true;
188
318
    case BPF_MODE_IND:
189
318
      CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
190
317
      return true;
191
1.26k
    }
192
17
    return false;
193
1.26k
  }
194
  /* LDX */
195
822
  if (BPF_MODE(bpf->op) == BPF_MODE_MEM) {
196
810
    CHECK_WRITABLE_AND_PUSH(ud, MI, bpf->dst);
197
809
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
198
808
    MCOperand_CreateImm0(MI, bpf->offset);
199
808
    return true;
200
809
  }
201
12
  return false;
202
822
}
203
204
static bool decodeStore(MCInst *MI, bpf_internal *bpf)
205
1.19k
{
206
  /* in cBPF, only BPF_ST* | BPF_MEM | BPF_W is valid
207
   * while in eBPF:
208
   * - BPF_STX | BPF_XADD | BPF_{W,DW}
209
   * - BPF_ST* | BPF_MEM | BPF_{W,H,B,DW}
210
   * are valid
211
   */
212
1.19k
  if (!EBPF_MODE(MI->csh->mode)) {
213
    /* can only store to M[] */
214
39
    if (bpf->op != (BPF_CLASS(bpf->op) | BPF_MODE_MEM | BPF_SIZE_W))
215
18
      return false;
216
21
    MCOperand_CreateImm0(MI, bpf->k);
217
21
    return true;
218
39
  }
219
220
  /* eBPF */
221
1.15k
  if (BPF_MODE(bpf->op) == BPF_MODE_ATOMIC) {
222
164
    if (BPF_CLASS(bpf->op) != BPF_CLASS_STX)
223
1
      return false;
224
163
    if (BPF_SIZE(bpf->op) != BPF_SIZE_W &&
225
56
        BPF_SIZE(bpf->op) != BPF_SIZE_DW)
226
0
      return false;
227
    /* xadd [dst + off], src */
228
163
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->dst);
229
161
    MCOperand_CreateImm0(MI, bpf->offset);
230
161
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
231
161
    return true;
232
161
  }
233
234
992
  if (BPF_MODE(bpf->op) != BPF_MODE_MEM)
235
21
    return false;
236
237
  /* st [dst + off], src */
238
971
  CHECK_READABLE_AND_PUSH(ud, MI, bpf->dst);
239
969
  MCOperand_CreateImm0(MI, bpf->offset);
240
969
  if (BPF_CLASS(bpf->op) == BPF_CLASS_ST)
241
511
    MCOperand_CreateImm0(MI, bpf->k);
242
458
  else
243
458
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
244
967
  return true;
245
969
}
246
247
static bool decodeALU(MCInst *MI, bpf_internal *bpf)
248
4.46k
{
249
  /* Set MI->Operands */
250
251
  /* cBPF */
252
4.46k
  if (!EBPF_MODE(MI->csh->mode)) {
253
1.71k
    if (BPF_OP(bpf->op) > BPF_ALU_XOR)
254
2
      return false;
255
    /* cBPF's NEG has no operands */
256
1.71k
    if (BPF_OP(bpf->op) == BPF_ALU_NEG)
257
241
      return true;
258
1.47k
    if (BPF_SRC(bpf->op) == BPF_SRC_K)
259
837
      MCOperand_CreateImm0(MI, bpf->k);
260
633
    else /* BPF_SRC_X */
261
633
      MCOperand_CreateReg0(MI, BPF_REG_X);
262
1.47k
    return true;
263
1.71k
  }
264
265
  /* eBPF */
266
267
2.75k
  if (BPF_OP(bpf->op) > BPF_ALU_END)
268
11
    return false;
269
  /* ENDian's imm must be one of 16, 32, 64 */
270
2.74k
  if (BPF_OP(bpf->op) == BPF_ALU_END) {
271
507
    if (bpf->k != 16 && bpf->k != 32 && bpf->k != 64)
272
35
      return false;
273
472
    if (BPF_CLASS(bpf->op) == BPF_CLASS_ALU64 &&
274
56
        BPF_SRC(bpf->op) != BPF_SRC_LITTLE)
275
1
      return false;
276
472
  }
277
278
  /* - op dst, imm
279
   * - op dst, src
280
   * - neg dst
281
   * - le<imm> dst
282
   */
283
  /* every ALU instructions have dst op */
284
2.70k
  CHECK_WRITABLE_AND_PUSH(ud, MI, bpf->dst);
285
286
  /* special cases */
287
2.69k
  if (BPF_OP(bpf->op) == BPF_ALU_NEG)
288
241
    return true;
289
2.45k
  if (BPF_OP(bpf->op) == BPF_ALU_END) {
290
    /* bpf->k must be one of 16, 32, 64 */
291
471
    bpf->op |= ((uint32_t)bpf->k << 4);
292
471
    return true;
293
471
  }
294
295
  /* normal cases */
296
1.98k
  if (BPF_SRC(bpf->op) == BPF_SRC_K) {
297
1.72k
    MCOperand_CreateImm0(MI, bpf->k);
298
1.72k
  } else { /* BPF_SRC_X */
299
256
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
300
256
  }
301
1.98k
  return true;
302
1.98k
}
303
304
static bool decodeJump(MCInst *MI, bpf_internal *bpf)
305
5.46k
{
306
  /* cBPF and eBPF are very different in class jump */
307
5.46k
  if (!EBPF_MODE(MI->csh->mode)) {
308
1.11k
    if (BPF_OP(bpf->op) > BPF_JUMP_JSET)
309
3
      return false;
310
311
    /* ja is a special case of jumps */
312
1.11k
    if (BPF_OP(bpf->op) == BPF_JUMP_JA) {
313
208
      MCOperand_CreateImm0(MI, bpf->k);
314
208
      return true;
315
208
    }
316
317
904
    if (BPF_SRC(bpf->op) == BPF_SRC_K)
318
348
      MCOperand_CreateImm0(MI, bpf->k);
319
556
    else /* BPF_SRC_X */
320
556
      MCOperand_CreateReg0(MI, BPF_REG_X);
321
904
    MCOperand_CreateImm0(MI, bpf->jt);
322
904
    MCOperand_CreateImm0(MI, bpf->jf);
323
4.35k
  } else {
324
4.35k
    if (BPF_OP(bpf->op) > BPF_JUMP_JSLE)
325
3
      return false;
326
327
    /* JMP32 has no CALL/EXIT instruction */
328
    /* No operands for exit */
329
4.34k
    if (BPF_OP(bpf->op) == BPF_JUMP_EXIT)
330
209
      return bpf->op == (BPF_CLASS_JMP | BPF_JUMP_EXIT);
331
4.14k
    if (BPF_OP(bpf->op) == BPF_JUMP_CALL) {
332
221
      if (bpf->op == (BPF_CLASS_JMP | BPF_JUMP_CALL)) {
333
196
        MCOperand_CreateImm0(MI, bpf->k);
334
196
        return true;
335
196
      }
336
25
      if (bpf->op ==
337
25
          (BPF_CLASS_JMP | BPF_JUMP_CALL | BPF_SRC_X)) {
338
24
        CHECK_READABLE_AND_PUSH(ud, MI, bpf->k);
339
10
        return true;
340
24
      }
341
1
      return false;
342
25
    }
343
344
    /* ja is a special case of jumps */
345
3.91k
    if (BPF_OP(bpf->op) == BPF_JUMP_JA) {
346
480
      if (BPF_SRC(bpf->op) != BPF_SRC_K)
347
3
        return false;
348
477
      if (BPF_CLASS(bpf->op) == BPF_CLASS_JMP)
349
89
        MCOperand_CreateImm0(MI, bpf->offset);
350
388
      else
351
388
        MCOperand_CreateImm0(MI, bpf->k);
352
353
477
      return true;
354
480
    }
355
356
    /* <j>  dst, src, +off */
357
3.43k
    CHECK_READABLE_AND_PUSH(ud, MI, bpf->dst);
358
3.43k
    if (BPF_SRC(bpf->op) == BPF_SRC_K)
359
3.32k
      MCOperand_CreateImm0(MI, bpf->k);
360
107
    else
361
107
      CHECK_READABLE_AND_PUSH(ud, MI, bpf->src);
362
3.42k
    MCOperand_CreateImm0(MI, bpf->offset);
363
3.42k
  }
364
4.33k
  return true;
365
5.46k
}
366
367
static bool decodeReturn(MCInst *MI, bpf_internal *bpf)
368
629
{
369
  /* Here only handles the BPF_RET class in cBPF */
370
629
  switch (BPF_RVAL(bpf->op)) {
371
216
  case BPF_SRC_K:
372
216
    MCOperand_CreateImm0(MI, bpf->k);
373
216
    return true;
374
201
  case BPF_SRC_X:
375
201
    MCOperand_CreateReg0(MI, BPF_REG_X);
376
201
    return true;
377
208
  case BPF_SRC_A:
378
208
    MCOperand_CreateReg0(MI, BPF_REG_A);
379
208
    return true;
380
629
  }
381
4
  return false;
382
629
}
383
384
static bool decodeMISC(MCInst *MI, bpf_internal *bpf)
385
105
{
386
105
  uint16_t op = bpf->op ^ BPF_CLASS_MISC;
387
105
  return op == BPF_MISCOP_TAX || op == BPF_MISCOP_TXA;
388
105
}
389
390
///< 1. Check if the instruction is valid
391
///< 2. Set MI->opcode
392
///< 3. Set MI->Operands
393
static bool getInstruction(MCInst *MI, bpf_internal *bpf)
394
16.9k
{
395
16.9k
  cs_detail *detail;
396
397
16.9k
  detail = MI->flat_insn->detail;
398
  // initialize detail
399
16.9k
  if (detail) {
400
16.9k
    memset(detail, 0, offsetof(cs_detail, bpf) + sizeof(cs_bpf));
401
16.9k
  }
402
403
16.9k
  MCInst_clear(MI);
404
405
16.9k
  switch (BPF_CLASS(bpf->op)) {
406
0
  default: /* should never happen */
407
0
    return false;
408
3.84k
  case BPF_CLASS_LD:
409
5.07k
  case BPF_CLASS_LDX:
410
5.07k
    return decodeLoad(MI, bpf);
411
547
  case BPF_CLASS_ST:
412
1.19k
  case BPF_CLASS_STX:
413
1.19k
    return decodeStore(MI, bpf);
414
2.88k
  case BPF_CLASS_ALU:
415
2.88k
    return decodeALU(MI, bpf);
416
2.95k
  case BPF_CLASS_JMP:
417
2.95k
    return decodeJump(MI, bpf);
418
3.13k
  case BPF_CLASS_RET:
419
    /* case BPF_CLASS_JMP32: */
420
3.13k
    if (EBPF_MODE(MI->csh->mode))
421
2.51k
      return decodeJump(MI, bpf);
422
629
    else
423
629
      return decodeReturn(MI, bpf);
424
1.68k
  case BPF_CLASS_MISC:
425
    /* case BPF_CLASS_ALU64: */
426
1.68k
    if (EBPF_MODE(MI->csh->mode))
427
1.58k
      return decodeALU(MI, bpf);
428
105
    else
429
105
      return decodeMISC(MI, bpf);
430
16.9k
  }
431
16.9k
}
432
433
// Check for regular load instructions
434
#define REG_LOAD_CASE(c) \
435
3.90k
  case BPF_SIZE_##c: \
436
3.90k
    if (BPF_CLASS(opcode) == BPF_CLASS_LD) \
437
3.90k
      return BPF_INS_LD##c; \
438
3.90k
    else \
439
3.90k
      return BPF_INS_LDX##c;
440
441
static bpf_insn op2insn_ld_cbpf(unsigned opcode)
442
3.90k
{
443
3.90k
  switch (BPF_SIZE(opcode)) {
444
2.27k
    REG_LOAD_CASE(W);
445
283
    REG_LOAD_CASE(H);
446
918
    REG_LOAD_CASE(B);
447
429
    REG_LOAD_CASE(DW);
448
3.90k
  }
449
450
0
  return BPF_INS_INVALID;
451
3.90k
}
452
#undef REG_LOAD_CASE
453
454
// Check for packet load instructions
455
#define PACKET_LOAD_CASE(c) \
456
1.08k
  case BPF_SIZE_##c: \
457
1.08k
    if (BPF_MODE(opcode) == BPF_MODE_ABS) \
458
1.08k
      return BPF_INS_LDABS##c; \
459
1.08k
    else if (BPF_MODE(opcode) == BPF_MODE_IND) \
460
290
      return BPF_INS_LDIND##c; \
461
290
    else \
462
290
      return BPF_INS_INVALID;
463
464
static bpf_insn op2insn_ld_ebpf(unsigned opcode)
465
2.01k
{
466
2.01k
  if (BPF_CLASS(opcode) == BPF_CLASS_LD) {
467
1.20k
    switch (BPF_SIZE(opcode)) {
468
418
      PACKET_LOAD_CASE(W);
469
414
      PACKET_LOAD_CASE(H);
470
251
      PACKET_LOAD_CASE(B);
471
1.20k
    }
472
1.20k
  }
473
474
  // If it's not a packet load instruction, it must be a regular load instruction
475
932
  return op2insn_ld_cbpf(opcode);
476
2.01k
}
477
#undef PACKET_LOAD_CASE
478
479
/* During parsing we already checked to make sure the size is D/DW and 
480
 * mode is STX and not ST, so we don't need to check again*/
481
#define ALU_CASE_REG(c) \
482
65
  case BPF_ALU_##c: \
483
65
    if (BPF_SIZE(opcode) == BPF_SIZE_W) \
484
65
      return BPF_INS_A##c; \
485
65
    else \
486
65
      return BPF_INS_A##c##64;
487
488
#define ALU_CASE_FETCH(c) \
489
70
  case BPF_ALU_##c | BPF_MODE_FETCH: \
490
70
    if (BPF_SIZE(opcode) == BPF_SIZE_W) \
491
70
      return BPF_INS_AF##c; \
492
70
    else \
493
70
      return BPF_INS_AF##c##64;
494
495
#define COMPLEX_CASE(c) \
496
13
  case BPF_ATOMIC_##c | BPF_MODE_FETCH: \
497
13
    if (BPF_SIZE(opcode) == BPF_SIZE_DW) \
498
13
      return BPF_INS_A##c##64; \
499
13
    else \
500
13
      return BPF_INS_INVALID;
501
502
#define CASE(c) \
503
988
  case BPF_SIZE_##c: \
504
988
    if (BPF_CLASS(opcode) == BPF_CLASS_ST) \
505
988
      return BPF_INS_ST##c; \
506
988
    else \
507
988
      return BPF_INS_STX##c;
508
509
static bpf_insn op2insn_st(unsigned opcode, const uint32_t imm)
510
1.14k
{
511
  /*
512
   * - BPF_STX | ALU atomic operations | BPF_{W,DW}
513
   * - BPF_STX | Complex atomic operations | BPF_{DW}
514
   * - BPF_ST* | BPF_MEM | BPF_{W,H,B,DW}
515
   */
516
517
1.14k
  if (BPF_MODE(opcode) == BPF_MODE_ATOMIC) {
518
161
    switch (imm) {
519
17
      ALU_CASE_REG(ADD);
520
4
      ALU_CASE_REG(OR);
521
9
      ALU_CASE_REG(AND);
522
35
      ALU_CASE_REG(XOR);
523
15
      ALU_CASE_FETCH(ADD);
524
15
      ALU_CASE_FETCH(OR);
525
24
      ALU_CASE_FETCH(AND);
526
16
      ALU_CASE_FETCH(XOR);
527
10
      COMPLEX_CASE(XCHG);
528
3
      COMPLEX_CASE(CMPXCHG);
529
13
    default: // Reached if complex atomic operation is used without fetch modifier
530
13
      return BPF_INS_INVALID;
531
161
    }
532
161
  }
533
534
  /* should be BPF_MEM */
535
988
  switch (BPF_SIZE(opcode)) {
536
345
    CASE(W);
537
239
    CASE(H);
538
136
    CASE(B);
539
268
    CASE(DW);
540
988
  }
541
542
0
  return BPF_INS_INVALID;
543
988
}
544
#undef CASE
545
546
#define CASE(c) \
547
3.16k
  case BPF_ALU_##c: \
548
3.16k
    CASE_IF(c)
549
550
#define CASE_IF(c) \
551
3.90k
  do { \
552
3.90k
    if (BPF_CLASS(opcode) == BPF_CLASS_ALU) \
553
3.90k
      return BPF_INS_##c; \
554
3.90k
    else \
555
3.90k
      return BPF_INS_##c##64; \
556
3.90k
  } while (0)
557
558
static bpf_insn op2insn_alu(unsigned opcode, const uint16_t off,
559
          const bool is_ebpf)
560
4.40k
{
561
  /* Endian is a special case */
562
4.40k
  if (BPF_OP(opcode) == BPF_ALU_END) {
563
471
    if (BPF_CLASS(opcode) == BPF_CLASS_ALU64) {
564
55
      switch (opcode ^ BPF_CLASS_ALU64 ^ BPF_ALU_END ^
565
55
        BPF_SRC_LITTLE) {
566
18
      case (16 << 4):
567
18
        return BPF_INS_BSWAP16;
568
23
      case (32 << 4):
569
23
        return BPF_INS_BSWAP32;
570
14
      case (64 << 4):
571
14
        return BPF_INS_BSWAP64;
572
0
      default:
573
0
        return BPF_INS_INVALID;
574
55
      }
575
55
    }
576
577
416
    switch (opcode ^ BPF_CLASS_ALU ^ BPF_ALU_END) {
578
13
    case BPF_SRC_LITTLE | (16 << 4):
579
13
      return BPF_INS_LE16;
580
215
    case BPF_SRC_LITTLE | (32 << 4):
581
215
      return BPF_INS_LE32;
582
28
    case BPF_SRC_LITTLE | (64 << 4):
583
28
      return BPF_INS_LE64;
584
30
    case BPF_SRC_BIG | (16 << 4):
585
30
      return BPF_INS_BE16;
586
50
    case BPF_SRC_BIG | (32 << 4):
587
50
      return BPF_INS_BE32;
588
80
    case BPF_SRC_BIG | (64 << 4):
589
80
      return BPF_INS_BE64;
590
416
    }
591
0
    return BPF_INS_INVALID;
592
416
  }
593
594
3.93k
  switch (BPF_OP(opcode)) {
595
182
    CASE(ADD);
596
95
    CASE(SUB);
597
273
    CASE(MUL);
598
443
    CASE(OR);
599
275
    CASE(AND);
600
153
    CASE(LSH);
601
476
    CASE(RSH);
602
482
    CASE(NEG);
603
330
    CASE(XOR);
604
460
    CASE(ARSH);
605
306
  case BPF_ALU_DIV:
606
306
    if (!is_ebpf || off == 0)
607
281
      CASE_IF(DIV);
608
25
    else if (off == 1)
609
17
      CASE_IF(SDIV);
610
8
    else
611
8
      return BPF_INS_INVALID;
612
354
  case BPF_ALU_MOD:
613
354
    if (!is_ebpf || off == 0)
614
305
      CASE_IF(MOD);
615
49
    else if (off == 1)
616
43
      CASE_IF(SMOD);
617
6
    else
618
6
      return BPF_INS_INVALID;
619
103
  case BPF_ALU_MOV:
620
    /* BPF_CLASS_ALU can have: mov, mov8s, mov16s
621
     * BPF_CLASS_ALU64 can have: mov, mov8s, mov16s, mov32s
622
     * */
623
103
    if (off == 0)
624
27
      CASE_IF(MOV);
625
76
    else if (off == 8)
626
14
      CASE_IF(MOVSB);
627
62
    else if (off == 16)
628
47
      CASE_IF(MOVSH);
629
15
    else if (off == 32 && BPF_CLASS(opcode) == BPF_CLASS_ALU64)
630
7
      return BPF_INS_MOVSW64;
631
8
    else
632
8
      return BPF_INS_INVALID;
633
3.93k
  }
634
635
0
  return BPF_INS_INVALID;
636
3.93k
}
637
#undef CASE_IF
638
#undef CASE
639
640
5.43k
#define BPF_CALLX (BPF_CLASS_JMP | BPF_JUMP_CALL | BPF_SRC_X)
641
642
#define CASE(c) \
643
4.33k
  case BPF_JUMP_##c: \
644
4.33k
    if (BPF_CLASS(opcode) == BPF_CLASS_JMP) \
645
4.33k
      return BPF_INS_##c; \
646
4.33k
    else \
647
4.33k
      return BPF_INS_##c##32;
648
649
#define SPEC_CASE(c) \
650
403
  case BPF_JUMP_##c: \
651
403
    if (BPF_CLASS(opcode) == BPF_CLASS_JMP) \
652
403
      return BPF_INS_##c; \
653
403
    else \
654
403
      return BPF_INS_INVALID;
655
656
static bpf_insn op2insn_jmp(unsigned opcode)
657
5.43k
{
658
5.43k
  if (opcode == BPF_CALLX) {
659
10
    return BPF_INS_CALLX;
660
10
  }
661
662
5.42k
  switch (BPF_OP(opcode)) {
663
685
  case BPF_JUMP_JA:
664
685
    if (BPF_CLASS(opcode) == BPF_CLASS_JMP)
665
297
      return BPF_INS_JA;
666
388
    else
667
388
      return BPF_INS_JAL;
668
365
    CASE(JEQ);
669
527
    CASE(JGT);
670
387
    CASE(JGE);
671
628
    CASE(JSET);
672
99
    CASE(JNE);
673
233
    CASE(JSGT);
674
542
    CASE(JSGE);
675
196
    SPEC_CASE(CALL);
676
207
    SPEC_CASE(EXIT);
677
398
    CASE(JLT);
678
514
    CASE(JLE);
679
360
    CASE(JSLT);
680
280
    CASE(JSLE);
681
5.42k
  }
682
683
0
  return BPF_INS_INVALID;
684
5.42k
}
685
#undef SPEC_CASE
686
#undef CASE
687
#undef BPF_CALLX
688
689
#ifndef CAPSTONE_DIET
690
691
static void update_regs_access(MCInst *MI, cs_detail *detail, bpf_insn insn_id,
692
             unsigned int opcode)
693
16.6k
{
694
16.6k
  if (insn_id == BPF_INS_INVALID)
695
0
    return;
696
  /*
697
   * In eBPF mode, only these instructions have implicit registers access:
698
   * - legacy ld{w,h,b,dw} * // w: r0
699
   * - exit // r: r0
700
   */
701
16.6k
  if (EBPF_MODE(MI->csh->mode)) {
702
10.1k
    switch (insn_id) {
703
8.70k
    default:
704
8.70k
      break;
705
8.70k
    case BPF_INS_LDABSW:
706
573
    case BPF_INS_LDABSH:
707
793
    case BPF_INS_LDABSB:
708
898
    case BPF_INS_LDINDW:
709
1.05k
    case BPF_INS_LDINDH:
710
1.08k
    case BPF_INS_LDINDB:
711
1.20k
    case BPF_INS_LDDW:
712
1.20k
      if (BPF_MODE(opcode) == BPF_MODE_ABS ||
713
407
          BPF_MODE(opcode) == BPF_MODE_IND)
714
1.11k
        map_add_implicit_write(MI, BPF_REG_R0);
715
1.20k
      break;
716
207
    case BPF_INS_EXIT:
717
207
      map_add_implicit_read(MI, BPF_REG_R0);
718
207
      break;
719
10.1k
    }
720
10.1k
    return;
721
10.1k
  }
722
723
  /* cBPF mode */
724
6.52k
  switch (BPF_CLASS(opcode)) {
725
625
  default:
726
625
    break;
727
2.57k
  case BPF_CLASS_LD:
728
2.57k
    map_add_implicit_write(MI, BPF_REG_A);
729
2.57k
    break;
730
400
  case BPF_CLASS_LDX:
731
400
    map_add_implicit_write(MI, BPF_REG_X);
732
400
    break;
733
18
  case BPF_CLASS_ST:
734
18
    map_add_implicit_read(MI, BPF_REG_A);
735
18
    break;
736
3
  case BPF_CLASS_STX:
737
3
    map_add_implicit_read(MI, BPF_REG_X);
738
3
    break;
739
1.71k
  case BPF_CLASS_ALU:
740
1.71k
    map_add_implicit_read(MI, BPF_REG_A);
741
1.71k
    map_add_implicit_write(MI, BPF_REG_A);
742
1.71k
    break;
743
1.11k
  case BPF_CLASS_JMP:
744
1.11k
    if (insn_id != BPF_INS_JA) // except the unconditional jump
745
904
      map_add_implicit_read(MI, BPF_REG_A);
746
1.11k
    break;
747
  /* case BPF_CLASS_RET: */
748
80
  case BPF_CLASS_MISC:
749
80
    if (insn_id == BPF_INS_TAX) {
750
13
      map_add_implicit_read(MI, BPF_REG_A);
751
13
      map_add_implicit_write(MI, BPF_REG_X);
752
67
    } else {
753
67
      map_add_implicit_read(MI, BPF_REG_X);
754
67
      map_add_implicit_write(MI, BPF_REG_A);
755
67
    }
756
80
    break;
757
6.52k
  }
758
6.52k
}
759
#endif
760
761
static bool setFinalOpcode(MCInst *MI, const bpf_internal *bpf)
762
16.6k
{
763
16.6k
  bpf_insn id = BPF_INS_INVALID;
764
16.6k
#ifndef CAPSTONE_DIET
765
16.6k
  cs_detail *detail;
766
767
16.6k
  detail = get_detail(MI);
768
16.6k
#endif
769
770
16.6k
  const uint16_t opcode = bpf->op;
771
16.6k
  switch (BPF_CLASS(opcode)) {
772
0
  default: // will never happen
773
0
    break;
774
3.77k
  case BPF_CLASS_LD:
775
4.98k
  case BPF_CLASS_LDX:
776
4.98k
    if (EBPF_MODE(MI->csh->mode))
777
2.01k
      id = op2insn_ld_ebpf(opcode);
778
2.97k
    else
779
2.97k
      id = op2insn_ld_cbpf(opcode);
780
4.98k
    add_group(MI, BPF_GRP_LOAD);
781
4.98k
    break;
782
529
  case BPF_CLASS_ST:
783
1.14k
  case BPF_CLASS_STX:
784
1.14k
    id = op2insn_st(opcode, bpf->k);
785
1.14k
    add_group(MI, BPF_GRP_STORE);
786
1.14k
    break;
787
2.85k
  case BPF_CLASS_ALU:
788
2.85k
    id = op2insn_alu(opcode, bpf->offset, EBPF_MODE(MI->csh->mode));
789
2.85k
    add_group(MI, BPF_GRP_ALU);
790
2.85k
    break;
791
2.93k
  case BPF_CLASS_JMP:
792
2.93k
    id = op2insn_jmp(opcode);
793
2.93k
#ifndef CAPSTONE_DIET
794
2.93k
    if (id == BPF_INS_CALL || id == BPF_INS_CALLX)
795
206
      add_group(MI, BPF_GRP_CALL);
796
2.72k
    else if (id == BPF_INS_EXIT)
797
207
      add_group(MI, BPF_GRP_RETURN);
798
2.51k
    else
799
2.51k
      add_group(MI, BPF_GRP_JUMP);
800
2.93k
#endif
801
2.93k
    break;
802
3.12k
  case BPF_CLASS_RET:
803
    /* case BPF_CLASS_JMP32: */
804
3.12k
    if (EBPF_MODE(MI->csh->mode)) {
805
2.49k
      id = op2insn_jmp(opcode);
806
2.49k
      add_group(MI, BPF_GRP_JUMP);
807
2.49k
    } else {
808
625
      id = BPF_INS_RET;
809
625
      add_group(MI, BPF_GRP_RETURN);
810
625
    }
811
3.12k
    break;
812
  // BPF_CLASS_MISC and BPF_CLASS_ALU64 have exactly same value
813
1.62k
  case BPF_CLASS_MISC:
814
    /* case BPF_CLASS_ALU64: */
815
1.62k
    if (EBPF_MODE(MI->csh->mode)) {
816
      // ALU64 in eBPF
817
1.54k
      id = op2insn_alu(opcode, bpf->offset, true);
818
1.54k
      add_group(MI, BPF_GRP_ALU);
819
1.54k
    } else {
820
80
      if (BPF_MISCOP(opcode) == BPF_MISCOP_TXA)
821
67
        id = BPF_INS_TXA;
822
13
      else
823
13
        id = BPF_INS_TAX;
824
80
      add_group(MI, BPF_GRP_MISC);
825
80
    }
826
1.62k
    break;
827
16.6k
  }
828
829
16.6k
  if (id == BPF_INS_INVALID)
830
36
    return false;
831
832
16.6k
  MCInst_setOpcodePub(MI, id);
833
16.6k
#undef PUSH_GROUP
834
835
16.6k
#ifndef CAPSTONE_DIET
836
16.6k
  if (detail) {
837
16.6k
    update_regs_access(MI, detail, id, opcode);
838
16.6k
  }
839
16.6k
#endif
840
16.6k
  return true;
841
16.6k
}
842
843
bool BPF_getInstruction(csh ud, const uint8_t *code, size_t code_len,
844
      MCInst *instr, uint16_t *size, uint64_t address,
845
      void *info)
846
17.2k
{
847
17.2k
  bpf_internal *bpf;
848
849
17.2k
  if (EBPF_MODE(instr->csh->mode))
850
10.5k
    bpf = fetch_ebpf(instr, code, code_len);
851
6.66k
  else
852
6.66k
    bpf = fetch_cbpf(instr, code, code_len);
853
17.2k
  if (bpf == NULL)
854
267
    return false;
855
16.9k
  if (!getInstruction(instr, bpf) || !setFinalOpcode(instr, bpf)) {
856
296
    cs_mem_free(bpf);
857
296
    return false;
858
296
  }
859
16.6k
  MCInst_setOpcode(instr, bpf->op);
860
861
16.6k
  *size = bpf->insn_size;
862
16.6k
  cs_mem_free(bpf);
863
864
  return true;
865
16.9k
}
866
867
#endif