1 /*
2 * Copyright 2000-2007 Niels Provos <provos@citi.umich.edu>
3 * Copyright 2007-2012 Niels Provos, Nick Mathewson
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
13 * 3. The name of the author may not be used to endorse or promote products
14 * derived from this software without specific prior written permission.
15 *
16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 */
27 #include "event2/event-config.h"
28 #include "evconfig-private.h"
29
30 #ifdef EVENT__HAVE_EPOLL
31
32 #include <stdint.h>
33 #include <sys/types.h>
34 #include <sys/resource.h>
35 #ifdef EVENT__HAVE_SYS_TIME_H
36 #include <sys/time.h>
37 #endif
38 #include <sys/queue.h>
39 #include <sys/epoll.h>
40 #include <signal.h>
41 #include <limits.h>
42 #include <stdio.h>
43 #include <stdlib.h>
44 #include <string.h>
45 #include <unistd.h>
46 #include <errno.h>
47 #ifdef EVENT__HAVE_FCNTL_H
48 #include <fcntl.h>
49 #endif
50 #ifdef EVENT__HAVE_SYS_TIMERFD_H
51 #include <sys/timerfd.h>
52 #endif
53
54 #include "event-internal.h"
55 #include "evsignal-internal.h"
56 #include "event2/thread.h"
57 #include "evthread-internal.h"
58 #include "log-internal.h"
59 #include "evmap-internal.h"
60 #include "changelist-internal.h"
61 #include "time-internal.h"
62
63 /* Since Linux 2.6.17, epoll is able to report about peer half-closed connection
64 using special EPOLLRDHUP flag on a read event.
65 */
66 #if !defined(EPOLLRDHUP)
67 #define EPOLLRDHUP 0
68 #define EARLY_CLOSE_IF_HAVE_RDHUP 0
69 #else
70 #define EARLY_CLOSE_IF_HAVE_RDHUP EV_FEATURE_EARLY_CLOSE
71 #endif
72
73 #include "epolltable-internal.h"
74
75 #if defined(EVENT__HAVE_SYS_TIMERFD_H) && \
76 defined(EVENT__HAVE_TIMERFD_CREATE) && \
77 defined(HAVE_POSIX_MONOTONIC) && defined(TFD_NONBLOCK) && \
78 defined(TFD_CLOEXEC)
79 /* Note that we only use timerfd if TFD_NONBLOCK and TFD_CLOEXEC are available
80 and working. This means that we can't support it on 2.6.25 (where timerfd
81 was introduced) or 2.6.26, since 2.6.27 introduced those flags.
82 */
83 #define USING_TIMERFD
84 #endif
85
86 struct epollop {
87 struct epoll_event *events;
88 int nevents;
89 int epfd;
90 #ifdef USING_TIMERFD
91 int timerfd;
92 #endif
93 };
94
95 static void *epoll_init(struct event_base *);
96 static int epoll_dispatch(struct event_base *, struct timeval *);
97 static void epoll_dealloc(struct event_base *);
98
99 static const struct eventop epollops_changelist = {
100 "epoll (with changelist)",
101 epoll_init,
102 event_changelist_add_,
103 event_changelist_del_,
104 epoll_dispatch,
105 epoll_dealloc,
106 1, /* need reinit */
107 EV_FEATURE_ET|EV_FEATURE_O1| EARLY_CLOSE_IF_HAVE_RDHUP,
108 EVENT_CHANGELIST_FDINFO_SIZE
109 };
110
111
112 static int epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
113 short old, short events, void *p);
114 static int epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
115 short old, short events, void *p);
116
117 const struct eventop epollops = {
118 "epoll",
119 epoll_init,
120 epoll_nochangelist_add,
121 epoll_nochangelist_del,
122 epoll_dispatch,
123 epoll_dealloc,
124 1, /* need reinit */
125 EV_FEATURE_ET|EV_FEATURE_O1|EV_FEATURE_EARLY_CLOSE,
126 0
127 };
128
129 #define INITIAL_NEVENT 32
130 #define MAX_NEVENT 4096
131
132 /* On Linux kernels at least up to 2.6.24.4, epoll can't handle timeout
133 * values bigger than (LONG_MAX - 999ULL)/HZ. HZ in the wild can be
134 * as big as 1000, and LONG_MAX can be as small as (1<<31)-1, so the
135 * largest number of msec we can support here is 2147482. Let's
136 * round that down by 47 seconds.
137 */
138 #define MAX_EPOLL_TIMEOUT_MSEC (35*60*1000)
139
140 static void *
epoll_init(struct event_base * base)141 epoll_init(struct event_base *base)
142 {
143 int epfd = -1;
144 struct epollop *epollop;
145
146 #ifdef EVENT__HAVE_EPOLL_CREATE1
147 /* First, try the shiny new epoll_create1 interface, if we have it. */
148 epfd = epoll_create1(EPOLL_CLOEXEC);
149 #endif
150 if (epfd == -1) {
151 /* Initialize the kernel queue using the old interface. (The
152 size field is ignored since 2.6.8.) */
153 if ((epfd = epoll_create(32000)) == -1) {
154 if (errno != ENOSYS)
155 event_warn("epoll_create");
156 return (NULL);
157 }
158 evutil_make_socket_closeonexec(epfd);
159 }
160
161 if (!(epollop = mm_calloc(1, sizeof(struct epollop)))) {
162 close(epfd);
163 return (NULL);
164 }
165
166 epollop->epfd = epfd;
167
168 /* Initialize fields */
169 epollop->events = mm_calloc(INITIAL_NEVENT, sizeof(struct epoll_event));
170 if (epollop->events == NULL) {
171 mm_free(epollop);
172 close(epfd);
173 return (NULL);
174 }
175 epollop->nevents = INITIAL_NEVENT;
176
177 if ((base->flags & EVENT_BASE_FLAG_EPOLL_USE_CHANGELIST) != 0 ||
178 ((base->flags & EVENT_BASE_FLAG_IGNORE_ENV) == 0 &&
179 evutil_getenv_("EVENT_EPOLL_USE_CHANGELIST") != NULL)) {
180
181 base->evsel = &epollops_changelist;
182 }
183
184 #ifdef USING_TIMERFD
185 /*
186 The epoll interface ordinarily gives us one-millisecond precision,
187 so on Linux it makes perfect sense to use the CLOCK_MONOTONIC_COARSE
188 timer. But when the user has set the new PRECISE_TIMER flag for an
189 event_base, we can try to use timerfd to give them finer granularity.
190 */
191 if ((base->flags & EVENT_BASE_FLAG_PRECISE_TIMER) &&
192 base->monotonic_timer.monotonic_clock == CLOCK_MONOTONIC) {
193 int fd;
194 fd = epollop->timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_NONBLOCK|TFD_CLOEXEC);
195 if (epollop->timerfd >= 0) {
196 struct epoll_event epev;
197 memset(&epev, 0, sizeof(epev));
198 epev.data.fd = epollop->timerfd;
199 epev.events = EPOLLIN;
200 if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, fd, &epev) < 0) {
201 event_warn("epoll_ctl(timerfd)");
202 close(fd);
203 epollop->timerfd = -1;
204 }
205 } else {
206 if (errno != EINVAL && errno != ENOSYS) {
207 /* These errors probably mean that we were
208 * compiled with timerfd/TFD_* support, but
209 * we're running on a kernel that lacks those.
210 */
211 event_warn("timerfd_create");
212 }
213 epollop->timerfd = -1;
214 }
215 } else {
216 epollop->timerfd = -1;
217 }
218 #endif
219
220 evsig_init_(base);
221
222 return (epollop);
223 }
224
225 static const char *
change_to_string(int change)226 change_to_string(int change)
227 {
228 change &= (EV_CHANGE_ADD|EV_CHANGE_DEL);
229 if (change == EV_CHANGE_ADD) {
230 return "add";
231 } else if (change == EV_CHANGE_DEL) {
232 return "del";
233 } else if (change == 0) {
234 return "none";
235 } else {
236 return "???";
237 }
238 }
239
240 static const char *
epoll_op_to_string(int op)241 epoll_op_to_string(int op)
242 {
243 return op == EPOLL_CTL_ADD?"ADD":
244 op == EPOLL_CTL_DEL?"DEL":
245 op == EPOLL_CTL_MOD?"MOD":
246 "???";
247 }
248
249 #define PRINT_CHANGES(op, events, ch, status) \
250 "Epoll %s(%d) on fd %d " status ". " \
251 "Old events were %d; " \
252 "read change was %d (%s); " \
253 "write change was %d (%s); " \
254 "close change was %d (%s)", \
255 epoll_op_to_string(op), \
256 events, \
257 ch->fd, \
258 ch->old_events, \
259 ch->read_change, \
260 change_to_string(ch->read_change), \
261 ch->write_change, \
262 change_to_string(ch->write_change), \
263 ch->close_change, \
264 change_to_string(ch->close_change)
265
266 static int
epoll_apply_one_change(struct event_base * base,struct epollop * epollop,const struct event_change * ch)267 epoll_apply_one_change(struct event_base *base,
268 struct epollop *epollop,
269 const struct event_change *ch)
270 {
271 struct epoll_event epev;
272 int op, events = 0;
273 int idx;
274
275 idx = EPOLL_OP_TABLE_INDEX(ch);
276 op = epoll_op_table[idx].op;
277 events = epoll_op_table[idx].events;
278
279 if (!events) {
280 EVUTIL_ASSERT(op == 0);
281 return 0;
282 }
283
284 if ((ch->read_change|ch->write_change|ch->close_change) & EV_CHANGE_ET)
285 events |= EPOLLET;
286
287 memset(&epev, 0, sizeof(epev));
288 epev.data.fd = ch->fd;
289 epev.events = events;
290 if (epoll_ctl(epollop->epfd, op, ch->fd, &epev) == 0) {
291 event_debug((PRINT_CHANGES(op, epev.events, ch, "okay")));
292 return 0;
293 }
294
295 switch (op) {
296 case EPOLL_CTL_MOD:
297 if (errno == ENOENT) {
298 /* If a MOD operation fails with ENOENT, the
299 * fd was probably closed and re-opened. We
300 * should retry the operation as an ADD.
301 */
302 if (epoll_ctl(epollop->epfd, EPOLL_CTL_ADD, ch->fd, &epev) == -1) {
303 event_warn("Epoll MOD(%d) on %d retried as ADD; that failed too",
304 (int)epev.events, ch->fd);
305 return -1;
306 } else {
307 event_debug(("Epoll MOD(%d) on %d retried as ADD; succeeded.",
308 (int)epev.events,
309 ch->fd));
310 return 0;
311 }
312 }
313 break;
314 case EPOLL_CTL_ADD:
315 if (errno == EEXIST) {
316 /* If an ADD operation fails with EEXIST,
317 * either the operation was redundant (as with a
318 * precautionary add), or we ran into a fun
319 * kernel bug where using dup*() to duplicate the
320 * same file into the same fd gives you the same epitem
321 * rather than a fresh one. For the second case,
322 * we must retry with MOD. */
323 if (epoll_ctl(epollop->epfd, EPOLL_CTL_MOD, ch->fd, &epev) == -1) {
324 event_warn("Epoll ADD(%d) on %d retried as MOD; that failed too",
325 (int)epev.events, ch->fd);
326 return -1;
327 } else {
328 event_debug(("Epoll ADD(%d) on %d retried as MOD; succeeded.",
329 (int)epev.events,
330 ch->fd));
331 return 0;
332 }
333 }
334 break;
335 case EPOLL_CTL_DEL:
336 if (errno == ENOENT || errno == EBADF || errno == EPERM) {
337 /* If a delete fails with one of these errors,
338 * that's fine too: we closed the fd before we
339 * got around to calling epoll_dispatch. */
340 event_debug(("Epoll DEL(%d) on fd %d gave %s: DEL was unnecessary.",
341 (int)epev.events,
342 ch->fd,
343 strerror(errno)));
344 return 0;
345 }
346 break;
347 default:
348 break;
349 }
350
351 event_warn(PRINT_CHANGES(op, epev.events, ch, "failed"));
352 return -1;
353 }
354
355 static int
epoll_apply_changes(struct event_base * base)356 epoll_apply_changes(struct event_base *base)
357 {
358 struct event_changelist *changelist = &base->changelist;
359 struct epollop *epollop = base->evbase;
360 struct event_change *ch;
361
362 int r = 0;
363 int i;
364
365 for (i = 0; i < changelist->n_changes; ++i) {
366 ch = &changelist->changes[i];
367 if (epoll_apply_one_change(base, epollop, ch) < 0)
368 r = -1;
369 }
370
371 return (r);
372 }
373
374 static int
epoll_nochangelist_add(struct event_base * base,evutil_socket_t fd,short old,short events,void * p)375 epoll_nochangelist_add(struct event_base *base, evutil_socket_t fd,
376 short old, short events, void *p)
377 {
378 struct event_change ch;
379 ch.fd = fd;
380 ch.old_events = old;
381 ch.read_change = ch.write_change = ch.close_change = 0;
382 if (events & EV_WRITE)
383 ch.write_change = EV_CHANGE_ADD |
384 (events & EV_ET);
385 if (events & EV_READ)
386 ch.read_change = EV_CHANGE_ADD |
387 (events & EV_ET);
388 if (events & EV_CLOSED)
389 ch.close_change = EV_CHANGE_ADD |
390 (events & EV_ET);
391
392 return epoll_apply_one_change(base, base->evbase, &ch);
393 }
394
395 static int
epoll_nochangelist_del(struct event_base * base,evutil_socket_t fd,short old,short events,void * p)396 epoll_nochangelist_del(struct event_base *base, evutil_socket_t fd,
397 short old, short events, void *p)
398 {
399 struct event_change ch;
400 ch.fd = fd;
401 ch.old_events = old;
402 ch.read_change = ch.write_change = ch.close_change = 0;
403 if (events & EV_WRITE)
404 ch.write_change = EV_CHANGE_DEL |
405 (events & EV_ET);
406 if (events & EV_READ)
407 ch.read_change = EV_CHANGE_DEL |
408 (events & EV_ET);
409 if (events & EV_CLOSED)
410 ch.close_change = EV_CHANGE_DEL |
411 (events & EV_ET);
412
413 return epoll_apply_one_change(base, base->evbase, &ch);
414 }
415
416 static int
epoll_dispatch(struct event_base * base,struct timeval * tv)417 epoll_dispatch(struct event_base *base, struct timeval *tv)
418 {
419 struct epollop *epollop = base->evbase;
420 struct epoll_event *events = epollop->events;
421 int i, res;
422 long timeout = -1;
423
424 #ifdef USING_TIMERFD
425 if (epollop->timerfd >= 0) {
426 struct itimerspec is;
427 is.it_interval.tv_sec = 0;
428 is.it_interval.tv_nsec = 0;
429 if (tv == NULL) {
430 /* No timeout; disarm the timer. */
431 is.it_value.tv_sec = 0;
432 is.it_value.tv_nsec = 0;
433 } else {
434 if (tv->tv_sec == 0 && tv->tv_usec == 0) {
435 /* we need to exit immediately; timerfd can't
436 * do that. */
437 timeout = 0;
438 }
439 is.it_value.tv_sec = tv->tv_sec;
440 is.it_value.tv_nsec = tv->tv_usec * 1000;
441 }
442 /* TODO: we could avoid unnecessary syscalls here by only
443 calling timerfd_settime when the top timeout changes, or
444 when we're called with a different timeval.
445 */
446 if (timerfd_settime(epollop->timerfd, 0, &is, NULL) < 0) {
447 event_warn("timerfd_settime");
448 }
449 } else
450 #endif
451 if (tv != NULL) {
452 timeout = evutil_tv_to_msec_(tv);
453 if (timeout < 0 || timeout > MAX_EPOLL_TIMEOUT_MSEC) {
454 /* Linux kernels can wait forever if the timeout is
455 * too big; see comment on MAX_EPOLL_TIMEOUT_MSEC. */
456 timeout = MAX_EPOLL_TIMEOUT_MSEC;
457 }
458 }
459
460 epoll_apply_changes(base);
461 event_changelist_remove_all_(&base->changelist, base);
462
463 EVBASE_RELEASE_LOCK(base, th_base_lock);
464
465 res = epoll_wait(epollop->epfd, events, epollop->nevents, timeout);
466
467 EVBASE_ACQUIRE_LOCK(base, th_base_lock);
468
469 if (res == -1) {
470 if (errno != EINTR) {
471 event_warn("epoll_wait");
472 return (-1);
473 }
474
475 return (0);
476 }
477
478 event_debug(("%s: epoll_wait reports %d", __func__, res));
479 EVUTIL_ASSERT(res <= epollop->nevents);
480
481 for (i = 0; i < res; i++) {
482 int what = events[i].events;
483 short ev = 0;
484 #ifdef USING_TIMERFD
485 if (events[i].data.fd == epollop->timerfd)
486 continue;
487 #endif
488
489 if (what & EPOLLERR) {
490 ev = EV_READ | EV_WRITE;
491 } else if ((what & EPOLLHUP) && !(what & EPOLLRDHUP)) {
492 ev = EV_READ | EV_WRITE;
493 } else {
494 if (what & EPOLLIN)
495 ev |= EV_READ;
496 if (what & EPOLLOUT)
497 ev |= EV_WRITE;
498 if (what & EPOLLRDHUP)
499 ev |= EV_CLOSED;
500 }
501
502 if (!ev)
503 continue;
504
505 evmap_io_active_(base, events[i].data.fd, ev | EV_ET);
506 }
507
508 if (res == epollop->nevents && epollop->nevents < MAX_NEVENT) {
509 /* We used all of the event space this time. We should
510 be ready for more events next time. */
511 int new_nevents = epollop->nevents * 2;
512 struct epoll_event *new_events;
513
514 new_events = mm_realloc(epollop->events,
515 new_nevents * sizeof(struct epoll_event));
516 if (new_events) {
517 epollop->events = new_events;
518 epollop->nevents = new_nevents;
519 }
520 }
521
522 return (0);
523 }
524
525
526 static void
epoll_dealloc(struct event_base * base)527 epoll_dealloc(struct event_base *base)
528 {
529 struct epollop *epollop = base->evbase;
530
531 evsig_dealloc_(base);
532 if (epollop->events)
533 mm_free(epollop->events);
534 if (epollop->epfd >= 0)
535 close(epollop->epfd);
536 #ifdef USING_TIMERFD
537 if (epollop->timerfd >= 0)
538 close(epollop->timerfd);
539 #endif
540
541 memset(epollop, 0, sizeof(struct epollop));
542 mm_free(epollop);
543 }
544
545 #endif /* EVENT__HAVE_EPOLL */
546