Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/kernel/locking/qspinlock_stat.h
25923 views
1
/* SPDX-License-Identifier: GPL-2.0-or-later */
2
/*
3
*
4
* Authors: Waiman Long <[email protected]>
5
*/
6
7
#include "lock_events.h"
8
9
#ifdef CONFIG_LOCK_EVENT_COUNTS
10
#ifdef CONFIG_PARAVIRT_SPINLOCKS
11
/*
12
* Collect pvqspinlock locking event counts
13
*/
14
#include <linux/sched.h>
15
#include <linux/sched/clock.h>
16
#include <linux/fs.h>
17
18
#define EVENT_COUNT(ev) lockevents[LOCKEVENT_ ## ev]
19
20
/*
21
* PV specific per-cpu counter
22
*/
23
static DEFINE_PER_CPU(u64, pv_kick_time);
24
25
/*
26
* Function to read and return the PV qspinlock counts.
27
*
28
* The following counters are handled specially:
29
* 1. pv_latency_kick
30
* Average kick latency (ns) = pv_latency_kick/pv_kick_unlock
31
* 2. pv_latency_wake
32
* Average wake latency (ns) = pv_latency_wake/pv_kick_wake
33
* 3. pv_hash_hops
34
* Average hops/hash = pv_hash_hops/pv_kick_unlock
35
*/
36
ssize_t lockevent_read(struct file *file, char __user *user_buf,
37
size_t count, loff_t *ppos)
38
{
39
char buf[64];
40
int cpu, id, len;
41
u64 sum = 0, kicks = 0;
42
43
/*
44
* Get the counter ID stored in file->f_inode->i_private
45
*/
46
id = (long)file_inode(file)->i_private;
47
48
if (id >= lockevent_num)
49
return -EBADF;
50
51
for_each_possible_cpu(cpu) {
52
sum += per_cpu(lockevents[id], cpu);
53
/*
54
* Need to sum additional counters for some of them
55
*/
56
switch (id) {
57
58
case LOCKEVENT_pv_latency_kick:
59
case LOCKEVENT_pv_hash_hops:
60
kicks += per_cpu(EVENT_COUNT(pv_kick_unlock), cpu);
61
break;
62
63
case LOCKEVENT_pv_latency_wake:
64
kicks += per_cpu(EVENT_COUNT(pv_kick_wake), cpu);
65
break;
66
}
67
}
68
69
if (id == LOCKEVENT_pv_hash_hops) {
70
u64 frac = 0;
71
72
if (kicks) {
73
frac = 100ULL * do_div(sum, kicks);
74
frac = DIV_ROUND_CLOSEST_ULL(frac, kicks);
75
}
76
77
/*
78
* Return a X.XX decimal number
79
*/
80
len = snprintf(buf, sizeof(buf) - 1, "%llu.%02llu\n",
81
sum, frac);
82
} else {
83
/*
84
* Round to the nearest ns
85
*/
86
if ((id == LOCKEVENT_pv_latency_kick) ||
87
(id == LOCKEVENT_pv_latency_wake)) {
88
if (kicks)
89
sum = DIV_ROUND_CLOSEST_ULL(sum, kicks);
90
}
91
len = snprintf(buf, sizeof(buf) - 1, "%llu\n", sum);
92
}
93
94
return simple_read_from_buffer(user_buf, count, ppos, buf, len);
95
}
96
97
/*
98
* PV hash hop count
99
*/
100
static inline void lockevent_pv_hop(int hopcnt)
101
{
102
this_cpu_add(EVENT_COUNT(pv_hash_hops), hopcnt);
103
}
104
105
/*
106
* Replacement function for pv_kick()
107
*/
108
static inline void __pv_kick(int cpu)
109
{
110
u64 start = sched_clock();
111
112
per_cpu(pv_kick_time, cpu) = start;
113
pv_kick(cpu);
114
this_cpu_add(EVENT_COUNT(pv_latency_kick), sched_clock() - start);
115
}
116
117
/*
118
* Replacement function for pv_wait()
119
*/
120
static inline void __pv_wait(u8 *ptr, u8 val)
121
{
122
u64 *pkick_time = this_cpu_ptr(&pv_kick_time);
123
124
*pkick_time = 0;
125
pv_wait(ptr, val);
126
if (*pkick_time) {
127
this_cpu_add(EVENT_COUNT(pv_latency_wake),
128
sched_clock() - *pkick_time);
129
lockevent_inc(pv_kick_wake);
130
}
131
}
132
133
#define pv_kick(c) __pv_kick(c)
134
#define pv_wait(p, v) __pv_wait(p, v)
135
136
#endif /* CONFIG_PARAVIRT_SPINLOCKS */
137
138
#else /* CONFIG_LOCK_EVENT_COUNTS */
139
140
static inline void lockevent_pv_hop(int hopcnt) { }
141
142
#endif /* CONFIG_LOCK_EVENT_COUNTS */
143
144