Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
awilliam
GitHub Repository: awilliam/linux-vfio
Path: blob/master/include/trace/events/sched.h
10820 views
1
#undef TRACE_SYSTEM
2
#define TRACE_SYSTEM sched
3
4
#if !defined(_TRACE_SCHED_H) || defined(TRACE_HEADER_MULTI_READ)
5
#define _TRACE_SCHED_H
6
7
#include <linux/sched.h>
8
#include <linux/tracepoint.h>
9
10
/*
11
* Tracepoint for calling kthread_stop, performed to end a kthread:
12
*/
13
TRACE_EVENT(sched_kthread_stop,
14
15
TP_PROTO(struct task_struct *t),
16
17
TP_ARGS(t),
18
19
TP_STRUCT__entry(
20
__array( char, comm, TASK_COMM_LEN )
21
__field( pid_t, pid )
22
),
23
24
TP_fast_assign(
25
memcpy(__entry->comm, t->comm, TASK_COMM_LEN);
26
__entry->pid = t->pid;
27
),
28
29
TP_printk("comm=%s pid=%d", __entry->comm, __entry->pid)
30
);
31
32
/*
33
* Tracepoint for the return value of the kthread stopping:
34
*/
35
TRACE_EVENT(sched_kthread_stop_ret,
36
37
TP_PROTO(int ret),
38
39
TP_ARGS(ret),
40
41
TP_STRUCT__entry(
42
__field( int, ret )
43
),
44
45
TP_fast_assign(
46
__entry->ret = ret;
47
),
48
49
TP_printk("ret=%d", __entry->ret)
50
);
51
52
/*
53
* Tracepoint for waking up a task:
54
*/
55
DECLARE_EVENT_CLASS(sched_wakeup_template,
56
57
TP_PROTO(struct task_struct *p, int success),
58
59
TP_ARGS(p, success),
60
61
TP_STRUCT__entry(
62
__array( char, comm, TASK_COMM_LEN )
63
__field( pid_t, pid )
64
__field( int, prio )
65
__field( int, success )
66
__field( int, target_cpu )
67
),
68
69
TP_fast_assign(
70
memcpy(__entry->comm, p->comm, TASK_COMM_LEN);
71
__entry->pid = p->pid;
72
__entry->prio = p->prio;
73
__entry->success = success;
74
__entry->target_cpu = task_cpu(p);
75
),
76
77
TP_printk("comm=%s pid=%d prio=%d success=%d target_cpu=%03d",
78
__entry->comm, __entry->pid, __entry->prio,
79
__entry->success, __entry->target_cpu)
80
);
81
82
DEFINE_EVENT(sched_wakeup_template, sched_wakeup,
83
TP_PROTO(struct task_struct *p, int success),
84
TP_ARGS(p, success));
85
86
/*
87
* Tracepoint for waking up a new task:
88
*/
89
DEFINE_EVENT(sched_wakeup_template, sched_wakeup_new,
90
TP_PROTO(struct task_struct *p, int success),
91
TP_ARGS(p, success));
92
93
#ifdef CREATE_TRACE_POINTS
94
static inline long __trace_sched_switch_state(struct task_struct *p)
95
{
96
long state = p->state;
97
98
#ifdef CONFIG_PREEMPT
99
/*
100
* For all intents and purposes a preempted task is a running task.
101
*/
102
if (task_thread_info(p)->preempt_count & PREEMPT_ACTIVE)
103
state = TASK_RUNNING;
104
#endif
105
106
return state;
107
}
108
#endif
109
110
/*
111
* Tracepoint for task switches, performed by the scheduler:
112
*/
113
TRACE_EVENT(sched_switch,
114
115
TP_PROTO(struct task_struct *prev,
116
struct task_struct *next),
117
118
TP_ARGS(prev, next),
119
120
TP_STRUCT__entry(
121
__array( char, prev_comm, TASK_COMM_LEN )
122
__field( pid_t, prev_pid )
123
__field( int, prev_prio )
124
__field( long, prev_state )
125
__array( char, next_comm, TASK_COMM_LEN )
126
__field( pid_t, next_pid )
127
__field( int, next_prio )
128
),
129
130
TP_fast_assign(
131
memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN);
132
__entry->prev_pid = prev->pid;
133
__entry->prev_prio = prev->prio;
134
__entry->prev_state = __trace_sched_switch_state(prev);
135
memcpy(__entry->prev_comm, prev->comm, TASK_COMM_LEN);
136
__entry->next_pid = next->pid;
137
__entry->next_prio = next->prio;
138
),
139
140
TP_printk("prev_comm=%s prev_pid=%d prev_prio=%d prev_state=%s ==> next_comm=%s next_pid=%d next_prio=%d",
141
__entry->prev_comm, __entry->prev_pid, __entry->prev_prio,
142
__entry->prev_state ?
143
__print_flags(__entry->prev_state, "|",
144
{ 1, "S"} , { 2, "D" }, { 4, "T" }, { 8, "t" },
145
{ 16, "Z" }, { 32, "X" }, { 64, "x" },
146
{ 128, "W" }) : "R",
147
__entry->next_comm, __entry->next_pid, __entry->next_prio)
148
);
149
150
/*
151
* Tracepoint for a task being migrated:
152
*/
153
TRACE_EVENT(sched_migrate_task,
154
155
TP_PROTO(struct task_struct *p, int dest_cpu),
156
157
TP_ARGS(p, dest_cpu),
158
159
TP_STRUCT__entry(
160
__array( char, comm, TASK_COMM_LEN )
161
__field( pid_t, pid )
162
__field( int, prio )
163
__field( int, orig_cpu )
164
__field( int, dest_cpu )
165
),
166
167
TP_fast_assign(
168
memcpy(__entry->comm, p->comm, TASK_COMM_LEN);
169
__entry->pid = p->pid;
170
__entry->prio = p->prio;
171
__entry->orig_cpu = task_cpu(p);
172
__entry->dest_cpu = dest_cpu;
173
),
174
175
TP_printk("comm=%s pid=%d prio=%d orig_cpu=%d dest_cpu=%d",
176
__entry->comm, __entry->pid, __entry->prio,
177
__entry->orig_cpu, __entry->dest_cpu)
178
);
179
180
DECLARE_EVENT_CLASS(sched_process_template,
181
182
TP_PROTO(struct task_struct *p),
183
184
TP_ARGS(p),
185
186
TP_STRUCT__entry(
187
__array( char, comm, TASK_COMM_LEN )
188
__field( pid_t, pid )
189
__field( int, prio )
190
),
191
192
TP_fast_assign(
193
memcpy(__entry->comm, p->comm, TASK_COMM_LEN);
194
__entry->pid = p->pid;
195
__entry->prio = p->prio;
196
),
197
198
TP_printk("comm=%s pid=%d prio=%d",
199
__entry->comm, __entry->pid, __entry->prio)
200
);
201
202
/*
203
* Tracepoint for freeing a task:
204
*/
205
DEFINE_EVENT(sched_process_template, sched_process_free,
206
TP_PROTO(struct task_struct *p),
207
TP_ARGS(p));
208
209
210
/*
211
* Tracepoint for a task exiting:
212
*/
213
DEFINE_EVENT(sched_process_template, sched_process_exit,
214
TP_PROTO(struct task_struct *p),
215
TP_ARGS(p));
216
217
/*
218
* Tracepoint for waiting on task to unschedule:
219
*/
220
DEFINE_EVENT(sched_process_template, sched_wait_task,
221
TP_PROTO(struct task_struct *p),
222
TP_ARGS(p));
223
224
/*
225
* Tracepoint for a waiting task:
226
*/
227
TRACE_EVENT(sched_process_wait,
228
229
TP_PROTO(struct pid *pid),
230
231
TP_ARGS(pid),
232
233
TP_STRUCT__entry(
234
__array( char, comm, TASK_COMM_LEN )
235
__field( pid_t, pid )
236
__field( int, prio )
237
),
238
239
TP_fast_assign(
240
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
241
__entry->pid = pid_nr(pid);
242
__entry->prio = current->prio;
243
),
244
245
TP_printk("comm=%s pid=%d prio=%d",
246
__entry->comm, __entry->pid, __entry->prio)
247
);
248
249
/*
250
* Tracepoint for do_fork:
251
*/
252
TRACE_EVENT(sched_process_fork,
253
254
TP_PROTO(struct task_struct *parent, struct task_struct *child),
255
256
TP_ARGS(parent, child),
257
258
TP_STRUCT__entry(
259
__array( char, parent_comm, TASK_COMM_LEN )
260
__field( pid_t, parent_pid )
261
__array( char, child_comm, TASK_COMM_LEN )
262
__field( pid_t, child_pid )
263
),
264
265
TP_fast_assign(
266
memcpy(__entry->parent_comm, parent->comm, TASK_COMM_LEN);
267
__entry->parent_pid = parent->pid;
268
memcpy(__entry->child_comm, child->comm, TASK_COMM_LEN);
269
__entry->child_pid = child->pid;
270
),
271
272
TP_printk("comm=%s pid=%d child_comm=%s child_pid=%d",
273
__entry->parent_comm, __entry->parent_pid,
274
__entry->child_comm, __entry->child_pid)
275
);
276
277
/*
278
* XXX the below sched_stat tracepoints only apply to SCHED_OTHER/BATCH/IDLE
279
* adding sched_stat support to SCHED_FIFO/RR would be welcome.
280
*/
281
DECLARE_EVENT_CLASS(sched_stat_template,
282
283
TP_PROTO(struct task_struct *tsk, u64 delay),
284
285
TP_ARGS(tsk, delay),
286
287
TP_STRUCT__entry(
288
__array( char, comm, TASK_COMM_LEN )
289
__field( pid_t, pid )
290
__field( u64, delay )
291
),
292
293
TP_fast_assign(
294
memcpy(__entry->comm, tsk->comm, TASK_COMM_LEN);
295
__entry->pid = tsk->pid;
296
__entry->delay = delay;
297
)
298
TP_perf_assign(
299
__perf_count(delay);
300
),
301
302
TP_printk("comm=%s pid=%d delay=%Lu [ns]",
303
__entry->comm, __entry->pid,
304
(unsigned long long)__entry->delay)
305
);
306
307
308
/*
309
* Tracepoint for accounting wait time (time the task is runnable
310
* but not actually running due to scheduler contention).
311
*/
312
DEFINE_EVENT(sched_stat_template, sched_stat_wait,
313
TP_PROTO(struct task_struct *tsk, u64 delay),
314
TP_ARGS(tsk, delay));
315
316
/*
317
* Tracepoint for accounting sleep time (time the task is not runnable,
318
* including iowait, see below).
319
*/
320
DEFINE_EVENT(sched_stat_template, sched_stat_sleep,
321
TP_PROTO(struct task_struct *tsk, u64 delay),
322
TP_ARGS(tsk, delay));
323
324
/*
325
* Tracepoint for accounting iowait time (time the task is not runnable
326
* due to waiting on IO to complete).
327
*/
328
DEFINE_EVENT(sched_stat_template, sched_stat_iowait,
329
TP_PROTO(struct task_struct *tsk, u64 delay),
330
TP_ARGS(tsk, delay));
331
332
/*
333
* Tracepoint for accounting runtime (time the task is executing
334
* on a CPU).
335
*/
336
TRACE_EVENT(sched_stat_runtime,
337
338
TP_PROTO(struct task_struct *tsk, u64 runtime, u64 vruntime),
339
340
TP_ARGS(tsk, runtime, vruntime),
341
342
TP_STRUCT__entry(
343
__array( char, comm, TASK_COMM_LEN )
344
__field( pid_t, pid )
345
__field( u64, runtime )
346
__field( u64, vruntime )
347
),
348
349
TP_fast_assign(
350
memcpy(__entry->comm, tsk->comm, TASK_COMM_LEN);
351
__entry->pid = tsk->pid;
352
__entry->runtime = runtime;
353
__entry->vruntime = vruntime;
354
)
355
TP_perf_assign(
356
__perf_count(runtime);
357
),
358
359
TP_printk("comm=%s pid=%d runtime=%Lu [ns] vruntime=%Lu [ns]",
360
__entry->comm, __entry->pid,
361
(unsigned long long)__entry->runtime,
362
(unsigned long long)__entry->vruntime)
363
);
364
365
/*
366
* Tracepoint for showing priority inheritance modifying a tasks
367
* priority.
368
*/
369
TRACE_EVENT(sched_pi_setprio,
370
371
TP_PROTO(struct task_struct *tsk, int newprio),
372
373
TP_ARGS(tsk, newprio),
374
375
TP_STRUCT__entry(
376
__array( char, comm, TASK_COMM_LEN )
377
__field( pid_t, pid )
378
__field( int, oldprio )
379
__field( int, newprio )
380
),
381
382
TP_fast_assign(
383
memcpy(__entry->comm, tsk->comm, TASK_COMM_LEN);
384
__entry->pid = tsk->pid;
385
__entry->oldprio = tsk->prio;
386
__entry->newprio = newprio;
387
),
388
389
TP_printk("comm=%s pid=%d oldprio=%d newprio=%d",
390
__entry->comm, __entry->pid,
391
__entry->oldprio, __entry->newprio)
392
);
393
394
#endif /* _TRACE_SCHED_H */
395
396
/* This part must be outside protection */
397
#include <trace/define_trace.h>
398
399