X-Git-Url: https://git.llucax.com/software/libev.git/blobdiff_plain/0b7b062fb0b1c4f1ab22a52bd91f3c86195a084e..5446fa15c54ea807294a3fe0645b588a9dae365e:/ev_kqueue.c?ds=sidebyside diff --git a/ev_kqueue.c b/ev_kqueue.c index 4933270..41a5dbe 100644 --- a/ev_kqueue.c +++ b/ev_kqueue.c @@ -39,14 +39,10 @@ kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) { struct kevent *ke; - array_needsize (struct kevent, kqueue_changes, kqueue_changemax, ++kqueue_changecnt, ); - - ke = &kqueue_changes [kqueue_changecnt - 1]; - memset (ke, 0, sizeof (struct kevent)); - ke->ident = fd; - ke->filter = filter; - ke->flags = flags; - ke->fflags = fflags; + ++kqueue_changecnt; + array_needsize (struct kevent, kqueue_changes, kqueue_changemax, kqueue_changecnt, EMPTY2); + + EV_SET (&kqueue_changes [kqueue_changecnt - 1], fd, filter, flags, fflags, 0, 0); } #ifndef NOTE_EOF @@ -56,17 +52,20 @@ kqueue_change (EV_P_ int fd, int filter, int flags, int fflags) static void kqueue_modify (EV_P_ int fd, int oev, int nev) { - /* to detect close/reopen reliably, we have to remove and re-add */ - /* event requests even when oev == nev */ + if (oev != nev) + { + if (oev & EV_READ) + kqueue_change (EV_A_ fd, EVFILT_READ , EV_DELETE, 0); - if (oev & EV_READ) - kqueue_change (EV_A_ fd, EVFILT_READ, EV_DELETE, 0); + if (oev & EV_WRITE) + kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); + } - if (oev & EV_WRITE) - kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_DELETE, 0); + /* to detect close/reopen reliably, we have to re-add */ + /* event requests even when oev == nev */ if (nev & EV_READ) - kqueue_change (EV_A_ fd, EVFILT_READ, EV_ADD, NOTE_EOF); + kqueue_change (EV_A_ fd, EVFILT_READ , EV_ADD, NOTE_EOF); if (nev & EV_WRITE) kqueue_change (EV_A_ fd, EVFILT_WRITE, EV_ADD, NOTE_EOF); @@ -83,12 +82,12 @@ kqueue_poll (EV_P_ ev_tstamp timeout) { ev_free (kqueue_events); kqueue_eventmax = array_roundsize (struct kevent, kqueue_changecnt); - kqueue_events = ev_malloc (sizeof (struct kevent) * kqueue_eventmax); + kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); } ts.tv_sec = (time_t)timeout; - ts.tv_nsec = (long)(timeout - (ev_tstamp)ts.tv_sec) * 1e9; - res = kevent (kqueue_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); + ts.tv_nsec = (long)((timeout - (ev_tstamp)ts.tv_sec) * 1e9); + res = kevent (backend_fd, kqueue_changes, kqueue_changecnt, kqueue_events, kqueue_eventmax, &ts); kqueue_changecnt = 0; if (res < 0) @@ -103,7 +102,7 @@ kqueue_poll (EV_P_ ev_tstamp timeout) { int fd = kqueue_events [i].ident; - if (kqueue_events [i].flags & EV_ERROR) + if (expect_false (kqueue_events [i].flags & EV_ERROR)) { int err = kqueue_events [i].data; @@ -147,7 +146,7 @@ kqueue_poll (EV_P_ ev_tstamp timeout) { ev_free (kqueue_events); kqueue_eventmax = array_roundsize (struct kevent, kqueue_eventmax << 1); - kqueue_events = ev_malloc (sizeof (struct kevent) * kqueue_eventmax); + kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); } } @@ -157,10 +156,10 @@ kqueue_init (EV_P_ int flags) struct kevent ch, ev; /* Initalize the kernel queue */ - if ((kqueue_fd = kqueue ()) < 0) + if ((backend_fd = kqueue ()) < 0) return 0; - fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ + fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* not sure if necessary, hopefully doesn't hurt */ /* Check for Mac OS X kqueue bug. */ ch.ident = -1; @@ -172,33 +171,33 @@ kqueue_init (EV_P_ int flags) * stick an error in ev. If kqueue is broken, then * kevent will fail. */ - if (kevent (kqueue_fd, &ch, 1, &ev, 1, 0) != 1 + if (kevent (backend_fd, &ch, 1, &ev, 1, 0) != 1 || ev.ident != -1 || ev.flags != EV_ERROR) { /* detected broken kqueue */ - close (kqueue_fd); + close (backend_fd); return 0; } - method_fudge = 1e-3; /* needed to compensate for kevent returning early */ - method_modify = kqueue_modify; - method_poll = kqueue_poll; + backend_fudge = 1e-3; /* needed to compensate for kevent returning early */ + backend_modify = kqueue_modify; + backend_poll = kqueue_poll; kqueue_eventmax = 64; /* intiial number of events receivable per poll */ - kqueue_events = ev_malloc (sizeof (struct kevent) * kqueue_eventmax); + kqueue_events = (struct kevent *)ev_malloc (sizeof (struct kevent) * kqueue_eventmax); kqueue_changes = 0; kqueue_changemax = 0; kqueue_changecnt = 0; - return EVMETHOD_KQUEUE; + return EVBACKEND_KQUEUE; } static void kqueue_destroy (EV_P) { - close (kqueue_fd); + close (backend_fd); ev_free (kqueue_events); ev_free (kqueue_changes); @@ -207,12 +206,12 @@ kqueue_destroy (EV_P) static void kqueue_fork (EV_P) { - close (kqueue_fd); + close (backend_fd); - while ((kqueue_fd = kqueue ()) < 0) + while ((backend_fd = kqueue ()) < 0) syserr ("(libev) kqueue"); - fcntl (kqueue_fd, F_SETFD, FD_CLOEXEC); + fcntl (backend_fd, F_SETFD, FD_CLOEXEC); /* re-register interest in fds */ fd_rearm_all (EV_A);