Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
freebsd
GitHub Repository: freebsd/freebsd-src
Path: blob/main/contrib/llvm-project/compiler-rt/lib/scudo/standalone/mutex.h
35291 views
1
//===-- mutex.h -------------------------------------------------*- C++ -*-===//
2
//
3
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4
// See https://llvm.org/LICENSE.txt for license information.
5
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6
//
7
//===----------------------------------------------------------------------===//
8
9
#ifndef SCUDO_MUTEX_H_
10
#define SCUDO_MUTEX_H_
11
12
#include "atomic_helpers.h"
13
#include "common.h"
14
#include "thread_annotations.h"
15
16
#include <string.h>
17
18
#if SCUDO_FUCHSIA
19
#include <lib/sync/mutex.h> // for sync_mutex_t
20
#endif
21
22
namespace scudo {
23
24
class CAPABILITY("mutex") HybridMutex {
25
public:
26
bool tryLock() TRY_ACQUIRE(true);
27
NOINLINE void lock() ACQUIRE() {
28
if (LIKELY(tryLock()))
29
return;
30
// The compiler may try to fully unroll the loop, ending up in a
31
// NumberOfTries*NumberOfYields block of pauses mixed with tryLocks. This
32
// is large, ugly and unneeded, a compact loop is better for our purpose
33
// here. Use a pragma to tell the compiler not to unroll the loop.
34
#ifdef __clang__
35
#pragma nounroll
36
#endif
37
for (u8 I = 0U; I < NumberOfTries; I++) {
38
delayLoop();
39
if (tryLock())
40
return;
41
}
42
lockSlow();
43
}
44
void unlock() RELEASE();
45
46
// TODO(chiahungduan): In general, we may want to assert the owner of lock as
47
// well. Given the current uses of HybridMutex, it's acceptable without
48
// asserting the owner. Re-evaluate this when we have certain scenarios which
49
// requires a more fine-grained lock granularity.
50
ALWAYS_INLINE void assertHeld() ASSERT_CAPABILITY(this) {
51
if (SCUDO_DEBUG)
52
assertHeldImpl();
53
}
54
55
private:
56
void delayLoop() {
57
// The value comes from the average time spent in accessing caches (which
58
// are the fastest operations) so that we are unlikely to wait too long for
59
// fast operations.
60
constexpr u32 SpinTimes = 16;
61
volatile u32 V = 0;
62
for (u32 I = 0; I < SpinTimes; ++I) {
63
u32 Tmp = V + 1;
64
V = Tmp;
65
}
66
}
67
68
void assertHeldImpl();
69
70
// TODO(chiahungduan): Adapt this value based on scenarios. E.g., primary and
71
// secondary allocator have different allocation times.
72
static constexpr u8 NumberOfTries = 32U;
73
74
#if SCUDO_LINUX
75
atomic_u32 M = {};
76
#elif SCUDO_FUCHSIA
77
sync_mutex_t M = {};
78
#endif
79
80
void lockSlow() ACQUIRE();
81
};
82
83
class SCOPED_CAPABILITY ScopedLock {
84
public:
85
explicit ScopedLock(HybridMutex &M) ACQUIRE(M) : Mutex(M) { Mutex.lock(); }
86
~ScopedLock() RELEASE() { Mutex.unlock(); }
87
88
private:
89
HybridMutex &Mutex;
90
91
ScopedLock(const ScopedLock &) = delete;
92
void operator=(const ScopedLock &) = delete;
93
};
94
95
} // namespace scudo
96
97
#endif // SCUDO_MUTEX_H_
98
99