Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
freebsd
GitHub Repository: freebsd/freebsd-src
Path: blob/main/contrib/llvm-project/compiler-rt/lib/scudo/standalone/fuchsia.cpp
35292 views
1
//===-- fuchsia.cpp ---------------------------------------------*- C++ -*-===//
2
//
3
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4
// See https://llvm.org/LICENSE.txt for license information.
5
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6
//
7
//===----------------------------------------------------------------------===//
8
9
#include "platform.h"
10
11
#if SCUDO_FUCHSIA
12
13
#include "common.h"
14
#include "mutex.h"
15
#include "string_utils.h"
16
17
#include <lib/sync/mutex.h> // for sync_mutex_t
18
#include <stdlib.h> // for getenv()
19
#include <zircon/compiler.h>
20
#include <zircon/process.h>
21
#include <zircon/sanitizer.h>
22
#include <zircon/status.h>
23
#include <zircon/syscalls.h>
24
25
namespace scudo {
26
27
uptr getPageSize() { return _zx_system_get_page_size(); }
28
29
void NORETURN die() { __builtin_trap(); }
30
31
// We zero-initialize the Extra parameter of map(), make sure this is consistent
32
// with ZX_HANDLE_INVALID.
33
static_assert(ZX_HANDLE_INVALID == 0, "");
34
35
static void NORETURN dieOnError(zx_status_t Status, const char *FnName,
36
uptr Size) {
37
ScopedString Error;
38
Error.append("SCUDO ERROR: %s failed with size %zuKB (%s)", FnName,
39
Size >> 10, zx_status_get_string(Status));
40
outputRaw(Error.data());
41
die();
42
}
43
44
static void *allocateVmar(uptr Size, MapPlatformData *Data, bool AllowNoMem) {
45
// Only scenario so far.
46
DCHECK(Data);
47
DCHECK_EQ(Data->Vmar, ZX_HANDLE_INVALID);
48
49
const zx_status_t Status = _zx_vmar_allocate(
50
_zx_vmar_root_self(),
51
ZX_VM_CAN_MAP_READ | ZX_VM_CAN_MAP_WRITE | ZX_VM_CAN_MAP_SPECIFIC, 0,
52
Size, &Data->Vmar, &Data->VmarBase);
53
if (UNLIKELY(Status != ZX_OK)) {
54
if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
55
dieOnError(Status, "zx_vmar_allocate", Size);
56
return nullptr;
57
}
58
return reinterpret_cast<void *>(Data->VmarBase);
59
}
60
61
void *map(void *Addr, uptr Size, const char *Name, uptr Flags,
62
MapPlatformData *Data) {
63
DCHECK_EQ(Size % getPageSizeCached(), 0);
64
const bool AllowNoMem = !!(Flags & MAP_ALLOWNOMEM);
65
66
// For MAP_NOACCESS, just allocate a Vmar and return.
67
if (Flags & MAP_NOACCESS)
68
return allocateVmar(Size, Data, AllowNoMem);
69
70
const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID)
71
? Data->Vmar
72
: _zx_vmar_root_self();
73
74
zx_status_t Status;
75
zx_handle_t Vmo;
76
uint64_t VmoSize = 0;
77
if (Data && Data->Vmo != ZX_HANDLE_INVALID) {
78
// If a Vmo was specified, it's a resize operation.
79
CHECK(Addr);
80
DCHECK(Flags & MAP_RESIZABLE);
81
Vmo = Data->Vmo;
82
VmoSize = Data->VmoSize;
83
Status = _zx_vmo_set_size(Vmo, VmoSize + Size);
84
if (Status != ZX_OK) {
85
if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
86
dieOnError(Status, "zx_vmo_set_size", VmoSize + Size);
87
return nullptr;
88
}
89
} else {
90
// Otherwise, create a Vmo and set its name.
91
Status = _zx_vmo_create(Size, ZX_VMO_RESIZABLE, &Vmo);
92
if (UNLIKELY(Status != ZX_OK)) {
93
if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
94
dieOnError(Status, "zx_vmo_create", Size);
95
return nullptr;
96
}
97
_zx_object_set_property(Vmo, ZX_PROP_NAME, Name, strlen(Name));
98
}
99
100
uintptr_t P;
101
zx_vm_option_t MapFlags =
102
ZX_VM_PERM_READ | ZX_VM_PERM_WRITE | ZX_VM_ALLOW_FAULTS;
103
if (Addr)
104
DCHECK(Data);
105
const uint64_t Offset =
106
Addr ? reinterpret_cast<uintptr_t>(Addr) - Data->VmarBase : 0;
107
if (Offset)
108
MapFlags |= ZX_VM_SPECIFIC;
109
Status = _zx_vmar_map(Vmar, MapFlags, Offset, Vmo, VmoSize, Size, &P);
110
if (UNLIKELY(Status != ZX_OK)) {
111
if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
112
dieOnError(Status, "zx_vmar_map", Size);
113
return nullptr;
114
}
115
116
if (Flags & MAP_PRECOMMIT) {
117
Status = _zx_vmar_op_range(Vmar, ZX_VMAR_OP_COMMIT, P, Size,
118
/*buffer=*/nullptr, /*buffer_size=*/0);
119
}
120
121
// No need to track the Vmo if we don't intend on resizing it. Close it.
122
if (Flags & MAP_RESIZABLE) {
123
DCHECK(Data);
124
if (Data->Vmo == ZX_HANDLE_INVALID)
125
Data->Vmo = Vmo;
126
else
127
DCHECK_EQ(Data->Vmo, Vmo);
128
} else {
129
CHECK_EQ(_zx_handle_close(Vmo), ZX_OK);
130
}
131
if (UNLIKELY(Status != ZX_OK)) {
132
if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
133
dieOnError(Status, "zx_vmar_op_range", Size);
134
return nullptr;
135
}
136
137
if (Data)
138
Data->VmoSize += Size;
139
140
return reinterpret_cast<void *>(P);
141
}
142
143
void unmap(void *Addr, uptr Size, uptr Flags, MapPlatformData *Data) {
144
if (Flags & UNMAP_ALL) {
145
DCHECK_NE(Data, nullptr);
146
const zx_handle_t Vmar = Data->Vmar;
147
DCHECK_NE(Vmar, _zx_vmar_root_self());
148
// Destroying the vmar effectively unmaps the whole mapping.
149
CHECK_EQ(_zx_vmar_destroy(Vmar), ZX_OK);
150
CHECK_EQ(_zx_handle_close(Vmar), ZX_OK);
151
} else {
152
const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID)
153
? Data->Vmar
154
: _zx_vmar_root_self();
155
const zx_status_t Status =
156
_zx_vmar_unmap(Vmar, reinterpret_cast<uintptr_t>(Addr), Size);
157
if (UNLIKELY(Status != ZX_OK))
158
dieOnError(Status, "zx_vmar_unmap", Size);
159
}
160
if (Data) {
161
if (Data->Vmo != ZX_HANDLE_INVALID)
162
CHECK_EQ(_zx_handle_close(Data->Vmo), ZX_OK);
163
memset(Data, 0, sizeof(*Data));
164
}
165
}
166
167
void setMemoryPermission(UNUSED uptr Addr, UNUSED uptr Size, UNUSED uptr Flags,
168
UNUSED MapPlatformData *Data) {
169
const zx_vm_option_t Prot =
170
(Flags & MAP_NOACCESS) ? 0 : (ZX_VM_PERM_READ | ZX_VM_PERM_WRITE);
171
DCHECK(Data);
172
DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID);
173
const zx_status_t Status = _zx_vmar_protect(Data->Vmar, Prot, Addr, Size);
174
if (Status != ZX_OK)
175
dieOnError(Status, "zx_vmar_protect", Size);
176
}
177
178
void releasePagesToOS(UNUSED uptr BaseAddress, uptr Offset, uptr Size,
179
MapPlatformData *Data) {
180
// TODO: DCHECK the BaseAddress is consistent with the data in
181
// MapPlatformData.
182
DCHECK(Data);
183
DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID);
184
DCHECK_NE(Data->Vmo, ZX_HANDLE_INVALID);
185
const zx_status_t Status =
186
_zx_vmo_op_range(Data->Vmo, ZX_VMO_OP_DECOMMIT, Offset, Size, NULL, 0);
187
CHECK_EQ(Status, ZX_OK);
188
}
189
190
const char *getEnv(const char *Name) { return getenv(Name); }
191
192
// Note: we need to flag these methods with __TA_NO_THREAD_SAFETY_ANALYSIS
193
// because the Fuchsia implementation of sync_mutex_t has clang thread safety
194
// annotations. Were we to apply proper capability annotations to the top level
195
// HybridMutex class itself, they would not be needed. As it stands, the
196
// thread analysis thinks that we are locking the mutex and accidentally leaving
197
// it locked on the way out.
198
bool HybridMutex::tryLock() __TA_NO_THREAD_SAFETY_ANALYSIS {
199
// Size and alignment must be compatible between both types.
200
return sync_mutex_trylock(&M) == ZX_OK;
201
}
202
203
void HybridMutex::lockSlow() __TA_NO_THREAD_SAFETY_ANALYSIS {
204
sync_mutex_lock(&M);
205
}
206
207
void HybridMutex::unlock() __TA_NO_THREAD_SAFETY_ANALYSIS {
208
sync_mutex_unlock(&M);
209
}
210
211
void HybridMutex::assertHeldImpl() __TA_NO_THREAD_SAFETY_ANALYSIS {}
212
213
u64 getMonotonicTime() { return _zx_clock_get_monotonic(); }
214
u64 getMonotonicTimeFast() { return _zx_clock_get_monotonic(); }
215
216
u32 getNumberOfCPUs() { return _zx_system_get_num_cpus(); }
217
218
u32 getThreadID() { return 0; }
219
220
bool getRandom(void *Buffer, uptr Length, UNUSED bool Blocking) {
221
static_assert(MaxRandomLength <= ZX_CPRNG_DRAW_MAX_LEN, "");
222
if (UNLIKELY(!Buffer || !Length || Length > MaxRandomLength))
223
return false;
224
_zx_cprng_draw(Buffer, Length);
225
return true;
226
}
227
228
void outputRaw(const char *Buffer) {
229
__sanitizer_log_write(Buffer, strlen(Buffer));
230
}
231
232
void setAbortMessage(const char *Message) {}
233
234
} // namespace scudo
235
236
#endif // SCUDO_FUCHSIA
237
238