Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
freebsd
GitHub Repository: freebsd/freebsd-src
Path: blob/main/contrib/libevent/epoll.c
39475 views
1
/*
2
* Copyright 2000-2007 Niels Provos <[email protected]>
3
* Copyright 2007-2012 Niels Provos, Nick Mathewson
4
*
5
* Redistribution and use in source and binary forms, with or without
6
* modification, are permitted provided that the following conditions
7
* are met:
8
* 1. Redistributions of source code must retain the above copyright
9
* notice, this list of conditions and the following disclaimer.
10
* 2. Redistributions in binary form must reproduce the above copyright
11
* notice, this list of conditions and the following disclaimer in the
12
* documentation and/or other materials provided with the distribution.
13
* 3. The name of the author may not be used to endorse or promote products
14
* derived from this software without specific prior written permission.
15
*
16
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26
*/
27
#include "event2/event-config.h"
28
#include "evconfig-private.h"
29
30
#ifdef EVENT__HAVE_EPOLL
31
32
#include <stdint.h>
33
#include <sys/types.h>
34
#include <sys/resource.h>
35
#ifdef EVENT__HAVE_SYS_TIME_H
36
#include <sys/time.h>
37
#endif
38
#include <sys/queue.h>
39
#include <sys/epoll.h>
40
#include <signal.h>
41
#include <limits.h>
42
#include <stdio.h>
43
#include <stdlib.h>
44
#include <string.h>
45
#include <unistd.h>
46
#include <errno.h>
47
#ifdef EVENT__HAVE_FCNTL_H
48
#include <fcntl.h>
49
#endif
50
#ifdef EVENT__HAVE_SYS_TIMERFD_H
51
#include <sys/timerfd.h>
52
#endif
53
54
#include "event-internal.h"
55
#include "evsignal-internal.h"
56
#include "event2/thread.h"
57
#include "evthread-internal.h"
58
#include "log-internal.h"
59
#include "evmap-internal.h"
60
#include "changelist-internal.h"
61
#include "time-internal.h"
62
63
/* Since Linux 2.6.17, epoll is able to report about peer half-closed connection
64
using special EPOLLRDHUP flag on a read event.
65
*/
66
#if !defined(EPOLLRDHUP)
67
#define EPOLLRDHUP 0
68
#define EARLY_CLOSE_IF_HAVE_RDHUP 0
69
#else
70
#define EARLY_CLOSE_IF_HAVE_RDHUP EV_FEATURE_EARLY_CLOSE
71
#endif
72
73
#include "epolltable-internal.h"
74
75
#if defined(EVENT__HAVE_SYS_TIMERFD_H) && \
76
defined(EVENT__HAVE_TIMERFD_CREATE) && \
77
defined(HAVE_POSIX_MONOTONIC) && defined(TFD_NONBLOCK) && \
78
defined(TFD_CLOEXEC)
79
/* Note that we only use timerfd if TFD_NONBLOCK and TFD_CLOEXEC are available
80
and working. This means that we can't support it on 2.6.25 (where timerfd
81
was introduced) or 2.6.26, since 2.6.27 introduced those flags.
82
*/
83
#define USING_TIMERFD
84
#endif
85
86
struct epollop {
87
struct epoll_event *events;
88
int nevents;
89
int epfd;
90
#ifdef USING_TIMERFD
91
int timerfd;
92
#endif
93
};
94
95
static void *epoll_init(struct event_base *);
96
static int epoll_dispatch(struct event_base *, struct timeval *);
97
static void epoll_dealloc(struct event_base *);
98
99
static const struct eventop epollops_changelist = {
100
"epoll (with changelist)",
101
epoll_init,
102
event_changelist_add_,
103
event_changelist_del_,
104
epoll_dispatch,
105
epoll_dealloc,
106
1, /* need reinit */
107
EV_FEATURE_ET|EV_FEATURE_O1| EARLY_CLOSE_IF_HAVE_RDHUP,
108
EVENT_CHANGELIST_FDINFO_SIZE
109
};
110
111
112
static int epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
113
short old, short events, void *p);
114
static int epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
115
short old, short events, void *p);
116
117
const struct eventop epollops = {
118
"epoll",
119
epoll_init,
120
epoll_nochangelist_add,
121
epoll_nochangelist_del,
122
epoll_dispatch,
123
epoll_dealloc,
124
1, /* need reinit */
125
EV_FEATURE_ET|EV_FEATURE_O1|EV_FEATURE_EARLY_CLOSE,
126
0
127
};
128
129
#define INITIAL_NEVENT 32
130
#define MAX_NEVENT 4096
131
132
/* On Linux kernels at least up to 2.6.24.4, epoll can't handle timeout
133
* values bigger than (LONG_MAX - 999ULL)/HZ. HZ in the wild can be
134
* as big as 1000, and LONG_MAX can be as small as (1<<31)-1, so the
135
* largest number of msec we can support here is 2147482. Let's
136
* round that down by 47 seconds.
137
*/
138
#define MAX_EPOLL_TIMEOUT_MSEC (35*60*1000)
139
140
static void *
141
epoll_init(struct event_base *base)
142
{
143
int epfd = -1;
144
struct epollop *epollop;
145
146
#ifdef EVENT__HAVE_EPOLL_CREATE1
147
/* First, try the shiny new epoll_create1 interface, if we have it. */
148
epfd = epoll_create1(EPOLL_CLOEXEC);
149
#endif
150
if (epfd == -1) {
151
/* Initialize the kernel queue using the old interface. (The
152
size field is ignored since 2.6.8.) */
153
if ((epfd = epoll_create(32000)) == -1) {
154
if (errno != ENOSYS)
155
event_warn("epoll_create");
156
return (NULL);
157
}
158
evutil_make_socket_closeonexec(epfd);
159
}
160
161
if (!(epollop = mm_calloc(1, sizeof(struct epollop)))) {
162
close(epfd);
163
return (NULL);
164
}
165
166
epollop->epfd = epfd;
167
168
/* Initialize fields */
169
epollop->events = mm_calloc(INITIAL_NEVENT, sizeof(struct epoll_event));
170
if (epollop->events == NULL) {
171
mm_free(epollop);
172
close(epfd);
173
return (NULL);
174
}
175
epollop->nevents = INITIAL_NEVENT;
176
177
if ((base->flags & EVENT_BASE_FLAG_EPOLL_USE_CHANGELIST) != 0 ||
178
((base->flags & EVENT_BASE_FLAG_IGNORE_ENV) == 0 &&
179
evutil_getenv_("EVENT_EPOLL_USE_CHANGELIST") != NULL)) {
180
181
base->evsel = &epollops_changelist;
182
}
183
184
#ifdef USING_TIMERFD
185
/*
186
The epoll interface ordinarily gives us one-millisecond precision,
187
so on Linux it makes perfect sense to use the CLOCK_MONOTONIC_COARSE
188
timer. But when the user has set the new PRECISE_TIMER flag for an
189
event_base, we can try to use timerfd to give them finer granularity.
190
*/
191
if ((base->flags & EVENT_BASE_FLAG_PRECISE_TIMER) &&
192
base->monotonic_timer.monotonic_clock == CLOCK_MONOTONIC) {
193
int fd;
194
fd = epollop->timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_NONBLOCK|TFD_CLOEXEC);
195
if (epollop->timerfd >= 0) {
196
struct epoll_event epev;
197
memset(&epev, 0, sizeof(epev));
198
epev.data.fd = epollop->timerfd;
199
epev.events = EPOLLIN;
200
if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, fd, &epev) < 0) {
201
event_warn("epoll_ctl(timerfd)");
202
close(fd);
203
epollop->timerfd = -1;
204
}
205
} else {
206
if (errno != EINVAL && errno != ENOSYS) {
207
/* These errors probably mean that we were
208
* compiled with timerfd/TFD_* support, but
209
* we're running on a kernel that lacks those.
210
*/
211
event_warn("timerfd_create");
212
}
213
epollop->timerfd = -1;
214
}
215
} else {
216
epollop->timerfd = -1;
217
}
218
#endif
219
220
evsig_init_(base);
221
222
return (epollop);
223
}
224
225
static const char *
226
change_to_string(int change)
227
{
228
change &= (EV_CHANGE_ADD|EV_CHANGE_DEL);
229
if (change == EV_CHANGE_ADD) {
230
return "add";
231
} else if (change == EV_CHANGE_DEL) {
232
return "del";
233
} else if (change == 0) {
234
return "none";
235
} else {
236
return "???";
237
}
238
}
239
240
static const char *
241
epoll_op_to_string(int op)
242
{
243
return op == EPOLL_CTL_ADD?"ADD":
244
op == EPOLL_CTL_DEL?"DEL":
245
op == EPOLL_CTL_MOD?"MOD":
246
"???";
247
}
248
249
#define PRINT_CHANGES(op, events, ch, status) \
250
"Epoll %s(%d) on fd %d " status ". " \
251
"Old events were %d; " \
252
"read change was %d (%s); " \
253
"write change was %d (%s); " \
254
"close change was %d (%s)", \
255
epoll_op_to_string(op), \
256
events, \
257
ch->fd, \
258
ch->old_events, \
259
ch->read_change, \
260
change_to_string(ch->read_change), \
261
ch->write_change, \
262
change_to_string(ch->write_change), \
263
ch->close_change, \
264
change_to_string(ch->close_change)
265
266
static int
267
epoll_apply_one_change(struct event_base *base,
268
struct epollop *epollop,
269
const struct event_change *ch)
270
{
271
struct epoll_event epev;
272
int op, events = 0;
273
int idx;
274
275
idx = EPOLL_OP_TABLE_INDEX(ch);
276
op = epoll_op_table[idx].op;
277
events = epoll_op_table[idx].events;
278
279
if (!events) {
280
EVUTIL_ASSERT(op == 0);
281
return 0;
282
}
283
284
if ((ch->read_change|ch->write_change|ch->close_change) & EV_CHANGE_ET)
285
events |= EPOLLET;
286
287
memset(&epev, 0, sizeof(epev));
288
epev.data.fd = ch->fd;
289
epev.events = events;
290
if (epoll_ctl(epollop->epfd, op, ch->fd, &epev) == 0) {
291
event_debug((PRINT_CHANGES(op, epev.events, ch, "okay")));
292
return 0;
293
}
294
295
switch (op) {
296
case EPOLL_CTL_MOD:
297
if (errno == ENOENT) {
298
/* If a MOD operation fails with ENOENT, the
299
* fd was probably closed and re-opened. We
300
* should retry the operation as an ADD.
301
*/
302
if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, ch->fd, &epev) == -1) {
303
event_warn("Epoll MOD(%d) on %d retried as ADD; that failed too",
304
(int)epev.events, ch->fd);
305
return -1;
306
} else {
307
event_debug(("Epoll MOD(%d) on %d retried as ADD; succeeded.",
308
(int)epev.events,
309
ch->fd));
310
return 0;
311
}
312
}
313
break;
314
case EPOLL_CTL_ADD:
315
if (errno == EEXIST) {
316
/* If an ADD operation fails with EEXIST,
317
* either the operation was redundant (as with a
318
* precautionary add), or we ran into a fun
319
* kernel bug where using dup*() to duplicate the
320
* same file into the same fd gives you the same epitem
321
* rather than a fresh one. For the second case,
322
* we must retry with MOD. */
323
if (epoll_ctl(epollop->epfd, EPOLL_CTL_MOD, ch->fd, &epev) == -1) {
324
event_warn("Epoll ADD(%d) on %d retried as MOD; that failed too",
325
(int)epev.events, ch->fd);
326
return -1;
327
} else {
328
event_debug(("Epoll ADD(%d) on %d retried as MOD; succeeded.",
329
(int)epev.events,
330
ch->fd));
331
return 0;
332
}
333
}
334
break;
335
case EPOLL_CTL_DEL:
336
if (errno == ENOENT || errno == EBADF || errno == EPERM) {
337
/* If a delete fails with one of these errors,
338
* that's fine too: we closed the fd before we
339
* got around to calling epoll_dispatch. */
340
event_debug(("Epoll DEL(%d) on fd %d gave %s: DEL was unnecessary.",
341
(int)epev.events,
342
ch->fd,
343
strerror(errno)));
344
return 0;
345
}
346
break;
347
default:
348
break;
349
}
350
351
event_warn(PRINT_CHANGES(op, epev.events, ch, "failed"));
352
return -1;
353
}
354
355
static int
356
epoll_apply_changes(struct event_base *base)
357
{
358
struct event_changelist *changelist = &base->changelist;
359
struct epollop *epollop = base->evbase;
360
struct event_change *ch;
361
362
int r = 0;
363
int i;
364
365
for (i = 0; i < changelist->n_changes; ++i) {
366
ch = &changelist->changes[i];
367
if (epoll_apply_one_change(base, epollop, ch) < 0)
368
r = -1;
369
}
370
371
return (r);
372
}
373
374
static int
375
epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
376
short old, short events, void *p)
377
{
378
struct event_change ch;
379
ch.fd = fd;
380
ch.old_events = old;
381
ch.read_change = ch.write_change = ch.close_change = 0;
382
if (events & EV_WRITE)
383
ch.write_change = EV_CHANGE_ADD |
384
(events & EV_ET);
385
if (events & EV_READ)
386
ch.read_change = EV_CHANGE_ADD |
387
(events & EV_ET);
388
if (events & EV_CLOSED)
389
ch.close_change = EV_CHANGE_ADD |
390
(events & EV_ET);
391
392
return epoll_apply_one_change(base, base->evbase, &ch);
393
}
394
395
static int
396
epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
397
short old, short events, void *p)
398
{
399
struct event_change ch;
400
ch.fd = fd;
401
ch.old_events = old;
402
ch.read_change = ch.write_change = ch.close_change = 0;
403
if (events & EV_WRITE)
404
ch.write_change = EV_CHANGE_DEL |
405
(events & EV_ET);
406
if (events & EV_READ)
407
ch.read_change = EV_CHANGE_DEL |
408
(events & EV_ET);
409
if (events & EV_CLOSED)
410
ch.close_change = EV_CHANGE_DEL |
411
(events & EV_ET);
412
413
return epoll_apply_one_change(base, base->evbase, &ch);
414
}
415
416
static int
417
epoll_dispatch(struct event_base *base, struct timeval *tv)
418
{
419
struct epollop *epollop = base->evbase;
420
struct epoll_event *events = epollop->events;
421
int i, res;
422
long timeout = -1;
423
424
#ifdef USING_TIMERFD
425
if (epollop->timerfd >= 0) {
426
struct itimerspec is;
427
is.it_interval.tv_sec = 0;
428
is.it_interval.tv_nsec = 0;
429
if (tv == NULL) {
430
/* No timeout; disarm the timer. */
431
is.it_value.tv_sec = 0;
432
is.it_value.tv_nsec = 0;
433
} else {
434
if (tv->tv_sec == 0 && tv->tv_usec == 0) {
435
/* we need to exit immediately; timerfd can't
436
* do that. */
437
timeout = 0;
438
}
439
is.it_value.tv_sec = tv->tv_sec;
440
is.it_value.tv_nsec = tv->tv_usec * 1000;
441
}
442
/* TODO: we could avoid unnecessary syscalls here by only
443
calling timerfd_settime when the top timeout changes, or
444
when we're called with a different timeval.
445
*/
446
if (timerfd_settime(epollop->timerfd, 0, &is, NULL) < 0) {
447
event_warn("timerfd_settime");
448
}
449
} else
450
#endif
451
if (tv != NULL) {
452
timeout = evutil_tv_to_msec_(tv);
453
if (timeout < 0 || timeout > MAX_EPOLL_TIMEOUT_MSEC) {
454
/* Linux kernels can wait forever if the timeout is
455
* too big; see comment on MAX_EPOLL_TIMEOUT_MSEC. */
456
timeout = MAX_EPOLL_TIMEOUT_MSEC;
457
}
458
}
459
460
epoll_apply_changes(base);
461
event_changelist_remove_all_(&base->changelist, base);
462
463
EVBASE_RELEASE_LOCK(base, th_base_lock);
464
465
res = epoll_wait(epollop->epfd, events, epollop->nevents, timeout);
466
467
EVBASE_ACQUIRE_LOCK(base, th_base_lock);
468
469
if (res == -1) {
470
if (errno != EINTR) {
471
event_warn("epoll_wait");
472
return (-1);
473
}
474
475
return (0);
476
}
477
478
event_debug(("%s: epoll_wait reports %d", __func__, res));
479
EVUTIL_ASSERT(res <= epollop->nevents);
480
481
for (i = 0; i < res; i++) {
482
int what = events[i].events;
483
short ev = 0;
484
#ifdef USING_TIMERFD
485
if (events[i].data.fd == epollop->timerfd)
486
continue;
487
#endif
488
489
if (what & EPOLLERR) {
490
ev = EV_READ | EV_WRITE;
491
} else if ((what & EPOLLHUP) && !(what & EPOLLRDHUP)) {
492
ev = EV_READ | EV_WRITE;
493
} else {
494
if (what & EPOLLIN)
495
ev |= EV_READ;
496
if (what & EPOLLOUT)
497
ev |= EV_WRITE;
498
if (what & EPOLLRDHUP)
499
ev |= EV_CLOSED;
500
}
501
502
if (!ev)
503
continue;
504
505
evmap_io_active_(base, events[i].data.fd, ev | EV_ET);
506
}
507
508
if (res == epollop->nevents && epollop->nevents < MAX_NEVENT) {
509
/* We used all of the event space this time. We should
510
be ready for more events next time. */
511
int new_nevents = epollop->nevents * 2;
512
struct epoll_event *new_events;
513
514
new_events = mm_realloc(epollop->events,
515
new_nevents * sizeof(struct epoll_event));
516
if (new_events) {
517
epollop->events = new_events;
518
epollop->nevents = new_nevents;
519
}
520
}
521
522
return (0);
523
}
524
525
526
static void
527
epoll_dealloc(struct event_base *base)
528
{
529
struct epollop *epollop = base->evbase;
530
531
evsig_dealloc_(base);
532
if (epollop->events)
533
mm_free(epollop->events);
534
if (epollop->epfd >= 0)
535
close(epollop->epfd);
536
#ifdef USING_TIMERFD
537
if (epollop->timerfd >= 0)
538
close(epollop->timerfd);
539
#endif
540
541
memset(epollop, 0, sizeof(struct epollop));
542
mm_free(epollop);
543
}
544
545
#endif /* EVENT__HAVE_EPOLL */
546
547