Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/tools/testing/selftests/bpf/bpf_experimental.h
26285 views
1
#ifndef __BPF_EXPERIMENTAL__
2
#define __BPF_EXPERIMENTAL__
3
4
#include <vmlinux.h>
5
#include <bpf/bpf_tracing.h>
6
#include <bpf/bpf_helpers.h>
7
#include <bpf/bpf_core_read.h>
8
9
#define __contains(name, node) __attribute__((btf_decl_tag("contains:" #name ":" #node)))
10
11
/* Description
12
* Allocates an object of the type represented by 'local_type_id' in
13
* program BTF. User may use the bpf_core_type_id_local macro to pass the
14
* type ID of a struct in program BTF.
15
*
16
* The 'local_type_id' parameter must be a known constant.
17
* The 'meta' parameter is rewritten by the verifier, no need for BPF
18
* program to set it.
19
* Returns
20
* A pointer to an object of the type corresponding to the passed in
21
* 'local_type_id', or NULL on failure.
22
*/
23
extern void *bpf_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
24
25
/* Convenience macro to wrap over bpf_obj_new_impl */
26
#define bpf_obj_new(type) ((type *)bpf_obj_new_impl(bpf_core_type_id_local(type), NULL))
27
28
/* Description
29
* Free an allocated object. All fields of the object that require
30
* destruction will be destructed before the storage is freed.
31
*
32
* The 'meta' parameter is rewritten by the verifier, no need for BPF
33
* program to set it.
34
* Returns
35
* Void.
36
*/
37
extern void bpf_obj_drop_impl(void *kptr, void *meta) __ksym;
38
39
/* Convenience macro to wrap over bpf_obj_drop_impl */
40
#define bpf_obj_drop(kptr) bpf_obj_drop_impl(kptr, NULL)
41
42
/* Description
43
* Increment the refcount on a refcounted local kptr, turning the
44
* non-owning reference input into an owning reference in the process.
45
*
46
* The 'meta' parameter is rewritten by the verifier, no need for BPF
47
* program to set it.
48
* Returns
49
* An owning reference to the object pointed to by 'kptr'
50
*/
51
extern void *bpf_refcount_acquire_impl(void *kptr, void *meta) __ksym;
52
53
/* Convenience macro to wrap over bpf_refcount_acquire_impl */
54
#define bpf_refcount_acquire(kptr) bpf_refcount_acquire_impl(kptr, NULL)
55
56
/* Description
57
* Add a new entry to the beginning of the BPF linked list.
58
*
59
* The 'meta' and 'off' parameters are rewritten by the verifier, no need
60
* for BPF programs to set them
61
* Returns
62
* 0 if the node was successfully added
63
* -EINVAL if the node wasn't added because it's already in a list
64
*/
65
extern int bpf_list_push_front_impl(struct bpf_list_head *head,
66
struct bpf_list_node *node,
67
void *meta, __u64 off) __ksym;
68
69
/* Convenience macro to wrap over bpf_list_push_front_impl */
70
#define bpf_list_push_front(head, node) bpf_list_push_front_impl(head, node, NULL, 0)
71
72
/* Description
73
* Add a new entry to the end of the BPF linked list.
74
*
75
* The 'meta' and 'off' parameters are rewritten by the verifier, no need
76
* for BPF programs to set them
77
* Returns
78
* 0 if the node was successfully added
79
* -EINVAL if the node wasn't added because it's already in a list
80
*/
81
extern int bpf_list_push_back_impl(struct bpf_list_head *head,
82
struct bpf_list_node *node,
83
void *meta, __u64 off) __ksym;
84
85
/* Convenience macro to wrap over bpf_list_push_back_impl */
86
#define bpf_list_push_back(head, node) bpf_list_push_back_impl(head, node, NULL, 0)
87
88
/* Description
89
* Remove the entry at the beginning of the BPF linked list.
90
* Returns
91
* Pointer to bpf_list_node of deleted entry, or NULL if list is empty.
92
*/
93
extern struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head) __ksym;
94
95
/* Description
96
* Remove the entry at the end of the BPF linked list.
97
* Returns
98
* Pointer to bpf_list_node of deleted entry, or NULL if list is empty.
99
*/
100
extern struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head) __ksym;
101
102
/* Description
103
* Remove 'node' from rbtree with root 'root'
104
* Returns
105
* Pointer to the removed node, or NULL if 'root' didn't contain 'node'
106
*/
107
extern struct bpf_rb_node *bpf_rbtree_remove(struct bpf_rb_root *root,
108
struct bpf_rb_node *node) __ksym;
109
110
/* Description
111
* Add 'node' to rbtree with root 'root' using comparator 'less'
112
*
113
* The 'meta' and 'off' parameters are rewritten by the verifier, no need
114
* for BPF programs to set them
115
* Returns
116
* 0 if the node was successfully added
117
* -EINVAL if the node wasn't added because it's already in a tree
118
*/
119
extern int bpf_rbtree_add_impl(struct bpf_rb_root *root, struct bpf_rb_node *node,
120
bool (less)(struct bpf_rb_node *a, const struct bpf_rb_node *b),
121
void *meta, __u64 off) __ksym;
122
123
/* Convenience macro to wrap over bpf_rbtree_add_impl */
124
#define bpf_rbtree_add(head, node, less) bpf_rbtree_add_impl(head, node, less, NULL, 0)
125
126
/* Description
127
* Return the first (leftmost) node in input tree
128
* Returns
129
* Pointer to the node, which is _not_ removed from the tree. If the tree
130
* contains no nodes, returns NULL.
131
*/
132
extern struct bpf_rb_node *bpf_rbtree_first(struct bpf_rb_root *root) __ksym;
133
134
/* Description
135
* Allocates a percpu object of the type represented by 'local_type_id' in
136
* program BTF. User may use the bpf_core_type_id_local macro to pass the
137
* type ID of a struct in program BTF.
138
*
139
* The 'local_type_id' parameter must be a known constant.
140
* The 'meta' parameter is rewritten by the verifier, no need for BPF
141
* program to set it.
142
* Returns
143
* A pointer to a percpu object of the type corresponding to the passed in
144
* 'local_type_id', or NULL on failure.
145
*/
146
extern void *bpf_percpu_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
147
148
/* Convenience macro to wrap over bpf_percpu_obj_new_impl */
149
#define bpf_percpu_obj_new(type) ((type __percpu_kptr *)bpf_percpu_obj_new_impl(bpf_core_type_id_local(type), NULL))
150
151
/* Description
152
* Free an allocated percpu object. All fields of the object that require
153
* destruction will be destructed before the storage is freed.
154
*
155
* The 'meta' parameter is rewritten by the verifier, no need for BPF
156
* program to set it.
157
* Returns
158
* Void.
159
*/
160
extern void bpf_percpu_obj_drop_impl(void *kptr, void *meta) __ksym;
161
162
struct bpf_iter_task_vma;
163
164
extern int bpf_iter_task_vma_new(struct bpf_iter_task_vma *it,
165
struct task_struct *task,
166
__u64 addr) __ksym;
167
extern struct vm_area_struct *bpf_iter_task_vma_next(struct bpf_iter_task_vma *it) __ksym;
168
extern void bpf_iter_task_vma_destroy(struct bpf_iter_task_vma *it) __ksym;
169
170
/* Convenience macro to wrap over bpf_obj_drop_impl */
171
#define bpf_percpu_obj_drop(kptr) bpf_percpu_obj_drop_impl(kptr, NULL)
172
173
/* Description
174
* Throw a BPF exception from the program, immediately terminating its
175
* execution and unwinding the stack. The supplied 'cookie' parameter
176
* will be the return value of the program when an exception is thrown,
177
* and the default exception callback is used. Otherwise, if an exception
178
* callback is set using the '__exception_cb(callback)' declaration tag
179
* on the main program, the 'cookie' parameter will be the callback's only
180
* input argument.
181
*
182
* Thus, in case of default exception callback, 'cookie' is subjected to
183
* constraints on the program's return value (as with R0 on exit).
184
* Otherwise, the return value of the marked exception callback will be
185
* subjected to the same checks.
186
*
187
* Note that throwing an exception with lingering resources (locks,
188
* references, etc.) will lead to a verification error.
189
*
190
* Note that callbacks *cannot* call this helper.
191
* Returns
192
* Never.
193
* Throws
194
* An exception with the specified 'cookie' value.
195
*/
196
extern void bpf_throw(u64 cookie) __ksym;
197
198
/* Description
199
* Acquire a reference on the exe_file member field belonging to the
200
* mm_struct that is nested within the supplied task_struct. The supplied
201
* task_struct must be trusted/referenced.
202
* Returns
203
* A referenced file pointer pointing to the exe_file member field of the
204
* mm_struct nested in the supplied task_struct, or NULL.
205
*/
206
extern struct file *bpf_get_task_exe_file(struct task_struct *task) __ksym;
207
208
/* Description
209
* Release a reference on the supplied file. The supplied file must be
210
* acquired.
211
*/
212
extern void bpf_put_file(struct file *file) __ksym;
213
214
/* Description
215
* Resolve a pathname for the supplied path and store it in the supplied
216
* buffer. The supplied path must be trusted/referenced.
217
* Returns
218
* A positive integer corresponding to the length of the resolved pathname,
219
* including the NULL termination character, stored in the supplied
220
* buffer. On error, a negative integer is returned.
221
*/
222
extern int bpf_path_d_path(struct path *path, char *buf, size_t buf__sz) __ksym;
223
224
/* This macro must be used to mark the exception callback corresponding to the
225
* main program. For example:
226
*
227
* int exception_cb(u64 cookie) {
228
* return cookie;
229
* }
230
*
231
* SEC("tc")
232
* __exception_cb(exception_cb)
233
* int main_prog(struct __sk_buff *ctx) {
234
* ...
235
* return TC_ACT_OK;
236
* }
237
*
238
* Here, exception callback for the main program will be 'exception_cb'. Note
239
* that this attribute can only be used once, and multiple exception callbacks
240
* specified for the main program will lead to verification error.
241
*/
242
#define __exception_cb(name) __attribute__((btf_decl_tag("exception_callback:" #name)))
243
244
#define __bpf_assert_signed(x) _Generic((x), \
245
unsigned long: 0, \
246
unsigned long long: 0, \
247
signed long: 1, \
248
signed long long: 1 \
249
)
250
251
#define __bpf_assert_check(LHS, op, RHS) \
252
_Static_assert(sizeof(&(LHS)), "1st argument must be an lvalue expression"); \
253
_Static_assert(sizeof(LHS) == 8, "Only 8-byte integers are supported\n"); \
254
_Static_assert(__builtin_constant_p(__bpf_assert_signed(LHS)), "internal static assert"); \
255
_Static_assert(__builtin_constant_p((RHS)), "2nd argument must be a constant expression")
256
257
#define __bpf_assert(LHS, op, cons, RHS, VAL) \
258
({ \
259
(void)bpf_throw; \
260
asm volatile ("if %[lhs] " op " %[rhs] goto +2; r1 = %[value]; call bpf_throw" \
261
: : [lhs] "r"(LHS), [rhs] cons(RHS), [value] "ri"(VAL) : ); \
262
})
263
264
#define __bpf_assert_op_sign(LHS, op, cons, RHS, VAL, supp_sign) \
265
({ \
266
__bpf_assert_check(LHS, op, RHS); \
267
if (__bpf_assert_signed(LHS) && !(supp_sign)) \
268
__bpf_assert(LHS, "s" #op, cons, RHS, VAL); \
269
else \
270
__bpf_assert(LHS, #op, cons, RHS, VAL); \
271
})
272
273
#define __bpf_assert_op(LHS, op, RHS, VAL, supp_sign) \
274
({ \
275
if (sizeof(typeof(RHS)) == 8) { \
276
const typeof(RHS) rhs_var = (RHS); \
277
__bpf_assert_op_sign(LHS, op, "r", rhs_var, VAL, supp_sign); \
278
} else { \
279
__bpf_assert_op_sign(LHS, op, "i", RHS, VAL, supp_sign); \
280
} \
281
})
282
283
#define __cmp_cannot_be_signed(x) \
284
__builtin_strcmp(#x, "==") == 0 || __builtin_strcmp(#x, "!=") == 0 || \
285
__builtin_strcmp(#x, "&") == 0
286
287
#define __is_signed_type(type) (((type)(-1)) < (type)1)
288
289
#define __bpf_cmp(LHS, OP, PRED, RHS, DEFAULT) \
290
({ \
291
__label__ l_true; \
292
bool ret = DEFAULT; \
293
asm volatile goto("if %[lhs] " OP " %[rhs] goto %l[l_true]" \
294
:: [lhs] "r"((short)LHS), [rhs] PRED (RHS) :: l_true); \
295
ret = !DEFAULT; \
296
l_true: \
297
ret; \
298
})
299
300
/* C type conversions coupled with comparison operator are tricky.
301
* Make sure BPF program is compiled with -Wsign-compare then
302
* __lhs OP __rhs below will catch the mistake.
303
* Be aware that we check only __lhs to figure out the sign of compare.
304
*/
305
#define _bpf_cmp(LHS, OP, RHS, UNLIKELY) \
306
({ \
307
typeof(LHS) __lhs = (LHS); \
308
typeof(RHS) __rhs = (RHS); \
309
bool ret; \
310
_Static_assert(sizeof(&(LHS)), "1st argument must be an lvalue expression"); \
311
(void)(__lhs OP __rhs); \
312
if (__cmp_cannot_be_signed(OP) || !__is_signed_type(typeof(__lhs))) { \
313
if (sizeof(__rhs) == 8) \
314
/* "i" will truncate 64-bit constant into s32, \
315
* so we have to use extra register via "r". \
316
*/ \
317
ret = __bpf_cmp(__lhs, #OP, "r", __rhs, UNLIKELY); \
318
else \
319
ret = __bpf_cmp(__lhs, #OP, "ri", __rhs, UNLIKELY); \
320
} else { \
321
if (sizeof(__rhs) == 8) \
322
ret = __bpf_cmp(__lhs, "s"#OP, "r", __rhs, UNLIKELY); \
323
else \
324
ret = __bpf_cmp(__lhs, "s"#OP, "ri", __rhs, UNLIKELY); \
325
} \
326
ret; \
327
})
328
329
#ifndef bpf_cmp_unlikely
330
#define bpf_cmp_unlikely(LHS, OP, RHS) _bpf_cmp(LHS, OP, RHS, true)
331
#endif
332
333
#ifndef bpf_cmp_likely
334
#define bpf_cmp_likely(LHS, OP, RHS) \
335
({ \
336
bool ret = 0; \
337
if (__builtin_strcmp(#OP, "==") == 0) \
338
ret = _bpf_cmp(LHS, !=, RHS, false); \
339
else if (__builtin_strcmp(#OP, "!=") == 0) \
340
ret = _bpf_cmp(LHS, ==, RHS, false); \
341
else if (__builtin_strcmp(#OP, "<=") == 0) \
342
ret = _bpf_cmp(LHS, >, RHS, false); \
343
else if (__builtin_strcmp(#OP, "<") == 0) \
344
ret = _bpf_cmp(LHS, >=, RHS, false); \
345
else if (__builtin_strcmp(#OP, ">") == 0) \
346
ret = _bpf_cmp(LHS, <=, RHS, false); \
347
else if (__builtin_strcmp(#OP, ">=") == 0) \
348
ret = _bpf_cmp(LHS, <, RHS, false); \
349
else \
350
asm volatile("r0 " #OP " invalid compare"); \
351
ret; \
352
})
353
#endif
354
355
/*
356
* Note that cond_break can only be portably used in the body of a breakable
357
* construct, whereas can_loop can be used anywhere.
358
*/
359
#ifdef __BPF_FEATURE_MAY_GOTO
360
#define can_loop \
361
({ __label__ l_break, l_continue; \
362
bool ret = true; \
363
asm volatile goto("may_goto %l[l_break]" \
364
:::: l_break); \
365
goto l_continue; \
366
l_break: ret = false; \
367
l_continue:; \
368
ret; \
369
})
370
371
#define __cond_break(expr) \
372
({ __label__ l_break, l_continue; \
373
asm volatile goto("may_goto %l[l_break]" \
374
:::: l_break); \
375
goto l_continue; \
376
l_break: expr; \
377
l_continue:; \
378
})
379
#else
380
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
381
#define can_loop \
382
({ __label__ l_break, l_continue; \
383
bool ret = true; \
384
asm volatile goto("1:.byte 0xe5; \
385
.byte 0; \
386
.long ((%l[l_break] - 1b - 8) / 8) & 0xffff; \
387
.short 0" \
388
:::: l_break); \
389
goto l_continue; \
390
l_break: ret = false; \
391
l_continue:; \
392
ret; \
393
})
394
395
#define __cond_break(expr) \
396
({ __label__ l_break, l_continue; \
397
asm volatile goto("1:.byte 0xe5; \
398
.byte 0; \
399
.long ((%l[l_break] - 1b - 8) / 8) & 0xffff; \
400
.short 0" \
401
:::: l_break); \
402
goto l_continue; \
403
l_break: expr; \
404
l_continue:; \
405
})
406
#else
407
#define can_loop \
408
({ __label__ l_break, l_continue; \
409
bool ret = true; \
410
asm volatile goto("1:.byte 0xe5; \
411
.byte 0; \
412
.long (((%l[l_break] - 1b - 8) / 8) & 0xffff) << 16; \
413
.short 0" \
414
:::: l_break); \
415
goto l_continue; \
416
l_break: ret = false; \
417
l_continue:; \
418
ret; \
419
})
420
421
#define __cond_break(expr) \
422
({ __label__ l_break, l_continue; \
423
asm volatile goto("1:.byte 0xe5; \
424
.byte 0; \
425
.long (((%l[l_break] - 1b - 8) / 8) & 0xffff) << 16; \
426
.short 0" \
427
:::: l_break); \
428
goto l_continue; \
429
l_break: expr; \
430
l_continue:; \
431
})
432
#endif
433
#endif
434
435
#define cond_break __cond_break(break)
436
#define cond_break_label(label) __cond_break(goto label)
437
438
#ifndef bpf_nop_mov
439
#define bpf_nop_mov(var) \
440
asm volatile("%[reg]=%[reg]"::[reg]"r"((short)var))
441
#endif
442
443
/* emit instruction:
444
* rX = rX .off = BPF_ADDR_SPACE_CAST .imm32 = (dst_as << 16) | src_as
445
*/
446
#ifndef bpf_addr_space_cast
447
#define bpf_addr_space_cast(var, dst_as, src_as)\
448
asm volatile(".byte 0xBF; \
449
.ifc %[reg], r0; \
450
.byte 0x00; \
451
.endif; \
452
.ifc %[reg], r1; \
453
.byte 0x11; \
454
.endif; \
455
.ifc %[reg], r2; \
456
.byte 0x22; \
457
.endif; \
458
.ifc %[reg], r3; \
459
.byte 0x33; \
460
.endif; \
461
.ifc %[reg], r4; \
462
.byte 0x44; \
463
.endif; \
464
.ifc %[reg], r5; \
465
.byte 0x55; \
466
.endif; \
467
.ifc %[reg], r6; \
468
.byte 0x66; \
469
.endif; \
470
.ifc %[reg], r7; \
471
.byte 0x77; \
472
.endif; \
473
.ifc %[reg], r8; \
474
.byte 0x88; \
475
.endif; \
476
.ifc %[reg], r9; \
477
.byte 0x99; \
478
.endif; \
479
.short %[off]; \
480
.long %[as]" \
481
: [reg]"+r"(var) \
482
: [off]"i"(BPF_ADDR_SPACE_CAST) \
483
, [as]"i"((dst_as << 16) | src_as));
484
#endif
485
486
void bpf_preempt_disable(void) __weak __ksym;
487
void bpf_preempt_enable(void) __weak __ksym;
488
489
typedef struct {
490
} __bpf_preempt_t;
491
492
static inline __bpf_preempt_t __bpf_preempt_constructor(void)
493
{
494
__bpf_preempt_t ret = {};
495
496
bpf_preempt_disable();
497
return ret;
498
}
499
static inline void __bpf_preempt_destructor(__bpf_preempt_t *t)
500
{
501
bpf_preempt_enable();
502
}
503
#define bpf_guard_preempt() \
504
__bpf_preempt_t ___bpf_apply(preempt, __COUNTER__) \
505
__attribute__((__unused__, __cleanup__(__bpf_preempt_destructor))) = \
506
__bpf_preempt_constructor()
507
508
/* Description
509
* Assert that a conditional expression is true.
510
* Returns
511
* Void.
512
* Throws
513
* An exception with the value zero when the assertion fails.
514
*/
515
#define bpf_assert(cond) if (!(cond)) bpf_throw(0);
516
517
/* Description
518
* Assert that a conditional expression is true.
519
* Returns
520
* Void.
521
* Throws
522
* An exception with the specified value when the assertion fails.
523
*/
524
#define bpf_assert_with(cond, value) if (!(cond)) bpf_throw(value);
525
526
/* Description
527
* Assert that LHS is in the range [BEG, END] (inclusive of both). This
528
* statement updates the known bounds of LHS during verification. Note
529
* that both BEG and END must be constant values, and must fit within the
530
* data type of LHS.
531
* Returns
532
* Void.
533
* Throws
534
* An exception with the value zero when the assertion fails.
535
*/
536
#define bpf_assert_range(LHS, BEG, END) \
537
({ \
538
_Static_assert(BEG <= END, "BEG must be <= END"); \
539
barrier_var(LHS); \
540
__bpf_assert_op(LHS, >=, BEG, 0, false); \
541
__bpf_assert_op(LHS, <=, END, 0, false); \
542
})
543
544
/* Description
545
* Assert that LHS is in the range [BEG, END] (inclusive of both). This
546
* statement updates the known bounds of LHS during verification. Note
547
* that both BEG and END must be constant values, and must fit within the
548
* data type of LHS.
549
* Returns
550
* Void.
551
* Throws
552
* An exception with the specified value when the assertion fails.
553
*/
554
#define bpf_assert_range_with(LHS, BEG, END, value) \
555
({ \
556
_Static_assert(BEG <= END, "BEG must be <= END"); \
557
barrier_var(LHS); \
558
__bpf_assert_op(LHS, >=, BEG, value, false); \
559
__bpf_assert_op(LHS, <=, END, value, false); \
560
})
561
562
struct bpf_iter_css_task;
563
struct cgroup_subsys_state;
564
extern int bpf_iter_css_task_new(struct bpf_iter_css_task *it,
565
struct cgroup_subsys_state *css, unsigned int flags) __weak __ksym;
566
extern struct task_struct *bpf_iter_css_task_next(struct bpf_iter_css_task *it) __weak __ksym;
567
extern void bpf_iter_css_task_destroy(struct bpf_iter_css_task *it) __weak __ksym;
568
569
struct bpf_iter_task;
570
extern int bpf_iter_task_new(struct bpf_iter_task *it,
571
struct task_struct *task, unsigned int flags) __weak __ksym;
572
extern struct task_struct *bpf_iter_task_next(struct bpf_iter_task *it) __weak __ksym;
573
extern void bpf_iter_task_destroy(struct bpf_iter_task *it) __weak __ksym;
574
575
struct bpf_iter_css;
576
extern int bpf_iter_css_new(struct bpf_iter_css *it,
577
struct cgroup_subsys_state *start, unsigned int flags) __weak __ksym;
578
extern struct cgroup_subsys_state *bpf_iter_css_next(struct bpf_iter_css *it) __weak __ksym;
579
extern void bpf_iter_css_destroy(struct bpf_iter_css *it) __weak __ksym;
580
581
extern int bpf_wq_init(struct bpf_wq *wq, void *p__map, unsigned int flags) __weak __ksym;
582
extern int bpf_wq_start(struct bpf_wq *wq, unsigned int flags) __weak __ksym;
583
extern int bpf_wq_set_callback_impl(struct bpf_wq *wq,
584
int (callback_fn)(void *map, int *key, void *value),
585
unsigned int flags__k, void *aux__ign) __ksym;
586
#define bpf_wq_set_callback(timer, cb, flags) \
587
bpf_wq_set_callback_impl(timer, cb, flags, NULL)
588
589
struct bpf_iter_kmem_cache;
590
extern int bpf_iter_kmem_cache_new(struct bpf_iter_kmem_cache *it) __weak __ksym;
591
extern struct kmem_cache *bpf_iter_kmem_cache_next(struct bpf_iter_kmem_cache *it) __weak __ksym;
592
extern void bpf_iter_kmem_cache_destroy(struct bpf_iter_kmem_cache *it) __weak __ksym;
593
594
struct bpf_iter_dmabuf;
595
extern int bpf_iter_dmabuf_new(struct bpf_iter_dmabuf *it) __weak __ksym;
596
extern struct dma_buf *bpf_iter_dmabuf_next(struct bpf_iter_dmabuf *it) __weak __ksym;
597
extern void bpf_iter_dmabuf_destroy(struct bpf_iter_dmabuf *it) __weak __ksym;
598
599
extern int bpf_cgroup_read_xattr(struct cgroup *cgroup, const char *name__str,
600
struct bpf_dynptr *value_p) __weak __ksym;
601
602
#endif
603
604