Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/samples/seccomp/bpf-helper.h
26278 views
1
/* SPDX-License-Identifier: GPL-2.0 */
2
/*
3
* Example wrapper around BPF macros.
4
*
5
* Copyright (c) 2012 The Chromium OS Authors <[email protected]>
6
* Author: Will Drewry <[email protected]>
7
*
8
* The code may be used by anyone for any purpose,
9
* and can serve as a starting point for developing
10
* applications using prctl(PR_SET_SECCOMP, 2, ...).
11
*
12
* No guarantees are provided with respect to the correctness
13
* or functionality of this code.
14
*/
15
#ifndef __BPF_HELPER_H__
16
#define __BPF_HELPER_H__
17
18
#include <asm/bitsperlong.h> /* for __BITS_PER_LONG */
19
#include <endian.h>
20
#include <linux/filter.h>
21
#include <linux/seccomp.h> /* for seccomp_data */
22
#include <linux/types.h>
23
#include <linux/unistd.h>
24
#include <stddef.h>
25
26
#define BPF_LABELS_MAX 256
27
struct bpf_labels {
28
int count;
29
struct __bpf_label {
30
const char *label;
31
__u32 location;
32
} labels[BPF_LABELS_MAX];
33
};
34
35
int bpf_resolve_jumps(struct bpf_labels *labels,
36
struct sock_filter *filter, size_t count);
37
__u32 seccomp_bpf_label(struct bpf_labels *labels, const char *label);
38
void seccomp_bpf_print(struct sock_filter *filter, size_t count);
39
40
#define JUMP_JT 0xff
41
#define JUMP_JF 0xff
42
#define LABEL_JT 0xfe
43
#define LABEL_JF 0xfe
44
45
#define ALLOW \
46
BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_ALLOW)
47
#define DENY \
48
BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_KILL)
49
#define JUMP(labels, label) \
50
BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
51
JUMP_JT, JUMP_JF)
52
#define LABEL(labels, label) \
53
BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
54
LABEL_JT, LABEL_JF)
55
#define SYSCALL(nr, jt) \
56
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (nr), 0, 1), \
57
jt
58
59
/* Lame, but just an example */
60
#define FIND_LABEL(labels, label) seccomp_bpf_label((labels), #label)
61
62
#define EXPAND(...) __VA_ARGS__
63
64
/* Ensure that we load the logically correct offset. */
65
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
66
#define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
67
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
68
#define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
69
#else
70
#error "Unknown endianness"
71
#endif
72
73
/* Map all width-sensitive operations */
74
#if __BITS_PER_LONG == 32
75
76
#define JEQ(x, jt) JEQ32(x, EXPAND(jt))
77
#define JNE(x, jt) JNE32(x, EXPAND(jt))
78
#define JGT(x, jt) JGT32(x, EXPAND(jt))
79
#define JLT(x, jt) JLT32(x, EXPAND(jt))
80
#define JGE(x, jt) JGE32(x, EXPAND(jt))
81
#define JLE(x, jt) JLE32(x, EXPAND(jt))
82
#define JA(x, jt) JA32(x, EXPAND(jt))
83
#define ARG(i) ARG_32(i)
84
85
#elif __BITS_PER_LONG == 64
86
87
/* Ensure that we load the logically correct offset. */
88
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
89
#define ENDIAN(_lo, _hi) _lo, _hi
90
#define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
91
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
92
#define ENDIAN(_lo, _hi) _hi, _lo
93
#define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
94
#endif
95
96
union arg64 {
97
struct {
98
__u32 ENDIAN(lo32, hi32);
99
};
100
__u64 u64;
101
};
102
103
#define JEQ(x, jt) \
104
JEQ64(((union arg64){.u64 = (x)}).lo32, \
105
((union arg64){.u64 = (x)}).hi32, \
106
EXPAND(jt))
107
#define JGT(x, jt) \
108
JGT64(((union arg64){.u64 = (x)}).lo32, \
109
((union arg64){.u64 = (x)}).hi32, \
110
EXPAND(jt))
111
#define JGE(x, jt) \
112
JGE64(((union arg64){.u64 = (x)}).lo32, \
113
((union arg64){.u64 = (x)}).hi32, \
114
EXPAND(jt))
115
#define JNE(x, jt) \
116
JNE64(((union arg64){.u64 = (x)}).lo32, \
117
((union arg64){.u64 = (x)}).hi32, \
118
EXPAND(jt))
119
#define JLT(x, jt) \
120
JLT64(((union arg64){.u64 = (x)}).lo32, \
121
((union arg64){.u64 = (x)}).hi32, \
122
EXPAND(jt))
123
#define JLE(x, jt) \
124
JLE64(((union arg64){.u64 = (x)}).lo32, \
125
((union arg64){.u64 = (x)}).hi32, \
126
EXPAND(jt))
127
128
#define JA(x, jt) \
129
JA64(((union arg64){.u64 = (x)}).lo32, \
130
((union arg64){.u64 = (x)}).hi32, \
131
EXPAND(jt))
132
#define ARG(i) ARG_64(i)
133
134
#else
135
#error __BITS_PER_LONG value unusable.
136
#endif
137
138
/* Loads the arg into A */
139
#define ARG_32(idx) \
140
BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx))
141
142
/* Loads lo into M[0] and hi into M[1] and A */
143
#define ARG_64(idx) \
144
BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx)), \
145
BPF_STMT(BPF_ST, 0), /* lo -> M[0] */ \
146
BPF_STMT(BPF_LD+BPF_W+BPF_ABS, HI_ARG(idx)), \
147
BPF_STMT(BPF_ST, 1) /* hi -> M[1] */
148
149
#define JEQ32(value, jt) \
150
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 0, 1), \
151
jt
152
153
#define JNE32(value, jt) \
154
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 1, 0), \
155
jt
156
157
#define JA32(value, jt) \
158
BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (value), 0, 1), \
159
jt
160
161
#define JGE32(value, jt) \
162
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 0, 1), \
163
jt
164
165
#define JGT32(value, jt) \
166
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 0, 1), \
167
jt
168
169
#define JLE32(value, jt) \
170
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 1, 0), \
171
jt
172
173
#define JLT32(value, jt) \
174
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 1, 0), \
175
jt
176
177
/*
178
* All the JXX64 checks assume lo is saved in M[0] and hi is saved in both
179
* A and M[1]. This invariant is kept by restoring A if necessary.
180
*/
181
#define JEQ64(lo, hi, jt) \
182
/* if (hi != arg.hi) goto NOMATCH; */ \
183
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
184
BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
185
/* if (lo != arg.lo) goto NOMATCH; */ \
186
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 0, 2), \
187
BPF_STMT(BPF_LD+BPF_MEM, 1), \
188
jt, \
189
BPF_STMT(BPF_LD+BPF_MEM, 1)
190
191
#define JNE64(lo, hi, jt) \
192
/* if (hi != arg.hi) goto MATCH; */ \
193
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 3), \
194
BPF_STMT(BPF_LD+BPF_MEM, 0), \
195
/* if (lo != arg.lo) goto MATCH; */ \
196
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 2, 0), \
197
BPF_STMT(BPF_LD+BPF_MEM, 1), \
198
jt, \
199
BPF_STMT(BPF_LD+BPF_MEM, 1)
200
201
#define JA64(lo, hi, jt) \
202
/* if (hi & arg.hi) goto MATCH; */ \
203
BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (hi), 3, 0), \
204
BPF_STMT(BPF_LD+BPF_MEM, 0), \
205
/* if (lo & arg.lo) goto MATCH; */ \
206
BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (lo), 0, 2), \
207
BPF_STMT(BPF_LD+BPF_MEM, 1), \
208
jt, \
209
BPF_STMT(BPF_LD+BPF_MEM, 1)
210
211
#define JGE64(lo, hi, jt) \
212
/* if (hi > arg.hi) goto MATCH; */ \
213
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
214
/* if (hi != arg.hi) goto NOMATCH; */ \
215
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
216
BPF_STMT(BPF_LD+BPF_MEM, 0), \
217
/* if (lo >= arg.lo) goto MATCH; */ \
218
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (lo), 0, 2), \
219
BPF_STMT(BPF_LD+BPF_MEM, 1), \
220
jt, \
221
BPF_STMT(BPF_LD+BPF_MEM, 1)
222
223
#define JGT64(lo, hi, jt) \
224
/* if (hi > arg.hi) goto MATCH; */ \
225
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
226
/* if (hi != arg.hi) goto NOMATCH; */ \
227
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
228
BPF_STMT(BPF_LD+BPF_MEM, 0), \
229
/* if (lo > arg.lo) goto MATCH; */ \
230
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 0, 2), \
231
BPF_STMT(BPF_LD+BPF_MEM, 1), \
232
jt, \
233
BPF_STMT(BPF_LD+BPF_MEM, 1)
234
235
#define JLE64(lo, hi, jt) \
236
/* if (hi < arg.hi) goto MATCH; */ \
237
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (hi), 0, 4), \
238
/* if (hi != arg.hi) goto NOMATCH; */ \
239
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
240
BPF_STMT(BPF_LD+BPF_MEM, 0), \
241
/* if (lo <= arg.lo) goto MATCH; */ \
242
BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 2, 0), \
243
BPF_STMT(BPF_LD+BPF_MEM, 1), \
244
jt, \
245
BPF_STMT(BPF_LD+BPF_MEM, 1)
246
247
#define JLT64(lo, hi, jt) \
248
/* if (hi < arg.hi) goto MATCH; */ \
249
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (hi), 0, 4), \
250
/* if (hi != arg.hi) goto NOMATCH; */ \
251
BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
252
BPF_STMT(BPF_LD+BPF_MEM, 0), \
253
/* if (lo < arg.lo) goto MATCH; */ \
254
BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (lo), 2, 0), \
255
BPF_STMT(BPF_LD+BPF_MEM, 1), \
256
jt, \
257
BPF_STMT(BPF_LD+BPF_MEM, 1)
258
259
#define LOAD_SYSCALL_NR \
260
BPF_STMT(BPF_LD+BPF_W+BPF_ABS, \
261
offsetof(struct seccomp_data, nr))
262
263
#endif /* __BPF_HELPER_H__ */
264
265