X-Git-Url: https://git.llucax.com/software/libev.git/blobdiff_plain/41a098dde97e50ab992166256ba6da1ffc2e431c..0a7b0ac074d9987c6fa23ab20be4842e4514c7b9:/ev.c diff --git a/ev.c b/ev.c index 0b2340e..8cb6141 100644 --- a/ev.c +++ b/ev.c @@ -34,7 +34,11 @@ extern "C" { #endif #ifndef EV_STANDALONE -# include "config.h" +# ifdef EV_CONFIG_H +# include EV_CONFIG_H +# else +# include "config.h" +# endif # if HAVE_CLOCK_GETTIME # ifndef EV_USE_MONOTONIC @@ -43,22 +47,61 @@ extern "C" { # ifndef EV_USE_REALTIME # define EV_USE_REALTIME 1 # endif +# else +# ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 0 +# endif +# ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 +# endif # endif -# if HAVE_SELECT && HAVE_SYS_SELECT_H && !defined (EV_USE_SELECT) -# define EV_USE_SELECT 1 +# ifndef EV_USE_SELECT +# if HAVE_SELECT && HAVE_SYS_SELECT_H +# define EV_USE_SELECT 1 +# else +# define EV_USE_SELECT 0 +# endif # endif -# if HAVE_POLL && HAVE_POLL_H && !defined (EV_USE_POLL) -# define EV_USE_POLL 1 +# ifndef EV_USE_POLL +# if HAVE_POLL && HAVE_POLL_H +# define EV_USE_POLL 1 +# else +# define EV_USE_POLL 0 +# endif # endif - -# if HAVE_EPOLL_CTL && HAVE_SYS_EPOLL_H && !defined (EV_USE_EPOLL) -# define EV_USE_EPOLL 1 + +# ifndef EV_USE_EPOLL +# if HAVE_EPOLL_CTL && HAVE_SYS_EPOLL_H +# define EV_USE_EPOLL 1 +# else +# define EV_USE_EPOLL 0 +# endif +# endif + +# ifndef EV_USE_KQUEUE +# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H +# define EV_USE_KQUEUE 1 +# else +# define EV_USE_KQUEUE 0 +# endif +# endif + +# ifndef EV_USE_PORT +# if HAVE_PORT_H && HAVE_PORT_CREATE +# define EV_USE_PORT 1 +# else +# define EV_USE_PORT 0 +# endif # endif -# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H && !defined (EV_USE_KQUEUE) -# define EV_USE_KQUEUE 1 +# ifndef EV_USE_INOTIFY +# if HAVE_INOTIFY_INIT && HAVE_SYS_INOTIFY_H +# define EV_USE_INOTIFY 1 +# else +# define EV_USE_INOTIFY 0 +# endif # endif #endif @@ -77,10 +120,16 @@ extern "C" { #include +#ifdef EV_H +# include EV_H +#else +# include "ev.h" +#endif + #ifndef _WIN32 -# include # include # include +# include #else # define WIN32_LEAN_AND_MEAN # include @@ -92,12 +141,15 @@ extern "C" { /**/ #ifndef EV_USE_MONOTONIC -# define EV_USE_MONOTONIC 1 +# define EV_USE_MONOTONIC 0 +#endif + +#ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 #endif #ifndef EV_USE_SELECT # define EV_USE_SELECT 1 -# define EV_SELECT_USE_FD_SET 1 #endif #ifndef EV_USE_POLL @@ -116,18 +168,32 @@ extern "C" { # define EV_USE_KQUEUE 0 #endif -#ifndef EV_USE_REALTIME -# define EV_USE_REALTIME 1 +#ifndef EV_USE_PORT +# define EV_USE_PORT 0 #endif -/**/ +#ifndef EV_USE_INOTIFY +# define EV_USE_INOTIFY 0 +#endif -/* darwin simply cannot be helped */ -#ifdef __APPLE__ -# undef EV_USE_POLL -# undef EV_USE_KQUEUE +#ifndef EV_PID_HASHSIZE +# if EV_MINIMAL +# define EV_PID_HASHSIZE 1 +# else +# define EV_PID_HASHSIZE 16 +# endif #endif +#ifndef EV_INOTIFY_HASHSIZE +# if EV_MINIMAL +# define EV_INOTIFY_HASHSIZE 1 +# else +# define EV_INOTIFY_HASHSIZE 16 +# endif +#endif + +/**/ + #ifndef CLOCK_MONOTONIC # undef EV_USE_MONOTONIC # define EV_USE_MONOTONIC 0 @@ -142,25 +208,35 @@ extern "C" { # include #endif +#if !EV_STAT_ENABLE +# define EV_USE_INOTIFY 0 +#endif + +#if EV_USE_INOTIFY +# include +#endif + /**/ #define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ -#define MAX_BLOCKTIME 59.731 /* never wait longer than this time (to detect time jumps) */ -#define PID_HASHSIZE 16 /* size of pid hash table, must be power of two */ -/*#define CLEANUP_INTERVAL 300. /* how often to try to free memory and re-check fds */ - -#ifdef EV_H -# include EV_H -#else -# include "ev.h" -#endif +#define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ +/*#define CLEANUP_INTERVAL (MAX_BLOCKTIME * 5.) /* how often to try to free memory and re-check fds */ #if __GNUC__ >= 3 # define expect(expr,value) __builtin_expect ((expr),(value)) -# define inline inline +# define inline_size static inline /* inline for codesize */ +# if EV_MINIMAL +# define noinline __attribute__ ((noinline)) +# define inline_speed static noinline +# else +# define noinline +# define inline_speed static inline +# endif #else # define expect(expr,value) (expr) -# define inline static +# define inline_speed static +# define inline_size static +# define noinline #endif #define expect_false(expr) expect ((expr) != 0, 0) @@ -172,9 +248,9 @@ extern "C" { #define EMPTY0 /* required for microsofts broken pseudo-c compiler */ #define EMPTY2(a,b) /* used to suppress some warnings */ -typedef struct ev_watcher *W; -typedef struct ev_watcher_list *WL; -typedef struct ev_watcher_time *WT; +typedef ev_watcher *W; +typedef ev_watcher_list *WL; +typedef ev_watcher_time *WT; static int have_monotonic; /* did clock_gettime (CLOCK_MONOTONIC) work? */ @@ -186,12 +262,13 @@ static int have_monotonic; /* did clock_gettime (CLOCK_MONOTONIC) work? */ static void (*syserr_cb)(const char *msg); -void ev_set_syserr_cb (void (*cb)(const char *msg)) +void +ev_set_syserr_cb (void (*cb)(const char *msg)) { syserr_cb = cb; } -static void +static void noinline syserr (const char *msg) { if (!msg) @@ -208,12 +285,13 @@ syserr (const char *msg) static void *(*alloc)(void *ptr, long size); -void ev_set_allocator (void *(*cb)(void *ptr, long size)) +void +ev_set_allocator (void *(*cb)(void *ptr, long size)) { alloc = cb; } -static void * +inline_speed void * ev_realloc (void *ptr, long size) { ptr = alloc ? alloc (ptr, size) : realloc (ptr, size); @@ -248,6 +326,13 @@ typedef struct int events; } ANPENDING; +#if EV_USE_INOTIFY +typedef struct +{ + WL head; +} ANFS; +#endif + #if EV_MULTIPLICITY struct ev_loop @@ -260,8 +345,8 @@ typedef struct }; #include "ev_wrap.h" - struct ev_loop default_loop_struct; - static struct ev_loop *default_loop; + static struct ev_loop default_loop_struct; + struct ev_loop *ev_default_loop_ptr; #else @@ -270,7 +355,7 @@ typedef struct #include "ev_vars.h" #undef VAR - static int default_loop; + static int ev_default_loop_ptr; #endif @@ -290,7 +375,7 @@ ev_time (void) #endif } -inline ev_tstamp +ev_tstamp inline_size get_clock (void) { #if EV_USE_MONOTONIC @@ -343,25 +428,12 @@ ev_now (EV_P) /*****************************************************************************/ -static void -anfds_init (ANFD *base, int count) -{ - while (count--) - { - base->head = 0; - base->events = EV_NONE; - base->reify = 0; - - ++base; - } -} - -void +void noinline ev_feed_event (EV_P_ void *w, int revents) { W w_ = (W)w; - if (w_->pending) + if (expect_false (w_->pending)) { pendings [ABSPRI (w_)][w_->pending - 1].events |= revents; return; @@ -373,7 +445,7 @@ ev_feed_event (EV_P_ void *w, int revents) pendings [ABSPRI (w_)][w_->pending - 1].events = revents; } -static void +void inline_size queue_events (EV_P_ W *events, int eventcnt, int type) { int i; @@ -382,13 +454,28 @@ queue_events (EV_P_ W *events, int eventcnt, int type) ev_feed_event (EV_A_ events [i], type); } -inline void +/*****************************************************************************/ + +void inline_size +anfds_init (ANFD *base, int count) +{ + while (count--) + { + base->head = 0; + base->events = EV_NONE; + base->reify = 0; + + ++base; + } +} + +void inline_speed fd_event (EV_P_ int fd, int revents) { ANFD *anfd = anfds + fd; - struct ev_io *w; + ev_io *w; - for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next) + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) { int ev = w->events & revents; @@ -403,9 +490,7 @@ ev_feed_fd_event (EV_P_ int fd, int revents) fd_event (EV_A_ fd, revents); } -/*****************************************************************************/ - -static void +void inline_size fd_reify (EV_P) { int i; @@ -414,11 +499,11 @@ fd_reify (EV_P) { int fd = fdchanges [i]; ANFD *anfd = anfds + fd; - struct ev_io *w; + ev_io *w; int events = 0; - for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next) + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) events |= w->events; #if EV_SELECT_IS_WINSOCKET @@ -432,17 +517,17 @@ fd_reify (EV_P) anfd->reify = 0; - method_modify (EV_A_ fd, anfd->events, events); + backend_modify (EV_A_ fd, anfd->events, events); anfd->events = events; } fdchangecnt = 0; } -static void +void inline_size fd_change (EV_P_ int fd) { - if (anfds [fd].reify) + if (expect_false (anfds [fd].reify)) return; anfds [fd].reify = 1; @@ -452,19 +537,19 @@ fd_change (EV_P_ int fd) fdchanges [fdchangecnt - 1] = fd; } -static void +void inline_speed fd_kill (EV_P_ int fd) { - struct ev_io *w; + ev_io *w; - while ((w = (struct ev_io *)anfds [fd].head)) + while ((w = (ev_io *)anfds [fd].head)) { ev_io_stop (EV_A_ w); ev_feed_event (EV_A_ (W)w, EV_ERROR | EV_READ | EV_WRITE); } } -static int +int inline_size fd_valid (int fd) { #ifdef _WIN32 @@ -475,7 +560,7 @@ fd_valid (int fd) } /* called on EBADF to verify fds */ -static void +static void noinline fd_ebadf (EV_P) { int fd; @@ -487,7 +572,7 @@ fd_ebadf (EV_P) } /* called on ENOMEM in select/poll to kill some fds and retry */ -static void +static void noinline fd_enomem (EV_P) { int fd; @@ -500,13 +585,12 @@ fd_enomem (EV_P) } } -/* usually called after fork if method needs to re-arm all fds from scratch */ -static void +/* usually called after fork if backend needs to re-arm all fds from scratch */ +static void noinline fd_rearm_all (EV_P) { int fd; - /* this should be highly optimised to not do anything but set a flag */ for (fd = 0; fd < anfdmax; ++fd) if (anfds [fd].events) { @@ -517,7 +601,7 @@ fd_rearm_all (EV_P) /*****************************************************************************/ -static void +void inline_speed upheap (WT *heap, int k) { WT w = heap [k]; @@ -534,7 +618,7 @@ upheap (WT *heap, int k) } -static void +void inline_speed downheap (WT *heap, int N, int k) { WT w = heap [k]; @@ -558,7 +642,7 @@ downheap (WT *heap, int N, int k) ((W)heap [k])->active = k + 1; } -inline void +void inline_size adjustheap (WT *heap, int N, int k) { upheap (heap, k); @@ -578,9 +662,9 @@ static int signalmax; static int sigpipe [2]; static sig_atomic_t volatile gotsig; -static struct ev_io sigev; +static ev_io sigev; -static void +void inline_size signals_init (ANSIG *base, int count) { while (count--) @@ -610,13 +694,13 @@ sighandler (int signum) } } -void +void noinline ev_feed_signal_event (EV_P_ int signum) { WL w; #if EV_MULTIPLICITY - assert (("feeding signal events is only supported in the default loop", loop == default_loop)); + assert (("feeding signal events is only supported in the default loop", loop == ev_default_loop_ptr)); #endif --signum; @@ -631,7 +715,7 @@ ev_feed_signal_event (EV_P_ int signum) } static void -sigcb (EV_P_ struct ev_io *iow, int revents) +sigcb (EV_P_ ev_io *iow, int revents) { int signum; @@ -643,7 +727,7 @@ sigcb (EV_P_ struct ev_io *iow, int revents) ev_feed_signal_event (EV_A_ signum + 1); } -inline void +void inline_size fd_intern (int fd) { #ifdef _WIN32 @@ -655,7 +739,7 @@ fd_intern (int fd) #endif } -static void +static void noinline siginit (EV_P) { fd_intern (sigpipe [0]); @@ -668,22 +752,18 @@ siginit (EV_P) /*****************************************************************************/ -static struct ev_child *childs [PID_HASHSIZE]; +static ev_child *childs [EV_PID_HASHSIZE]; #ifndef _WIN32 -static struct ev_signal childev; +static ev_signal childev; -#ifndef WCONTINUED -# define WCONTINUED 0 -#endif - -static void -child_reap (EV_P_ struct ev_signal *sw, int chain, int pid, int status) +void inline_speed +child_reap (EV_P_ ev_signal *sw, int chain, int pid, int status) { - struct ev_child *w; + ev_child *w; - for (w = (struct ev_child *)childs [chain & (PID_HASHSIZE - 1)]; w; w = (struct ev_child *)((WL)w)->next) + for (w = (ev_child *)childs [chain & (EV_PID_HASHSIZE - 1)]; w; w = (ev_child *)((WL)w)->next) if (w->pid == pid || !w->pid) { ev_priority (w) = ev_priority (sw); /* need to do it *now* */ @@ -693,25 +773,38 @@ child_reap (EV_P_ struct ev_signal *sw, int chain, int pid, int status) } } +#ifndef WCONTINUED +# define WCONTINUED 0 +#endif + static void -childcb (EV_P_ struct ev_signal *sw, int revents) +childcb (EV_P_ ev_signal *sw, int revents) { int pid, status; - if (0 < (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED))) - { - /* make sure we are called again until all childs have been reaped */ - ev_feed_event (EV_A_ (W)sw, EV_SIGNAL); + /* some systems define WCONTINUED but then fail to support it (linux 2.4) */ + if (0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED))) + if (!WCONTINUED + || errno != EINVAL + || 0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED))) + return; - child_reap (EV_A_ sw, pid, pid, status); - child_reap (EV_A_ sw, 0, pid, status); /* this might trigger a watcher twice, but event catches that */ - } + /* make sure we are called again until all childs have been reaped */ + /* we need to do it this way so that the callback gets called before we continue */ + ev_feed_event (EV_A_ (W)sw, EV_SIGNAL); + + child_reap (EV_A_ sw, pid, pid, status); + if (EV_PID_HASHSIZE > 1) + child_reap (EV_A_ sw, 0, pid, status); /* this might trigger a watcher twice, but feed_event catches that */ } #endif /*****************************************************************************/ +#if EV_USE_PORT +# include "ev_port.c" +#endif #if EV_USE_KQUEUE # include "ev_kqueue.c" #endif @@ -738,7 +831,7 @@ ev_version_minor (void) } /* return true if we are running with elevated privileges and should ignore env variables */ -static int +int inline_size enable_secure (void) { #ifdef _WIN32 @@ -750,15 +843,61 @@ enable_secure (void) } unsigned int -ev_method (EV_P) +ev_supported_backends (void) { - return method; + unsigned int flags = 0; + + if (EV_USE_PORT ) flags |= EVBACKEND_PORT; + if (EV_USE_KQUEUE) flags |= EVBACKEND_KQUEUE; + if (EV_USE_EPOLL ) flags |= EVBACKEND_EPOLL; + if (EV_USE_POLL ) flags |= EVBACKEND_POLL; + if (EV_USE_SELECT) flags |= EVBACKEND_SELECT; + + return flags; } -static void +unsigned int +ev_recommended_backends (void) +{ + unsigned int flags = ev_supported_backends (); + +#ifndef __NetBSD__ + /* kqueue is borked on everything but netbsd apparently */ + /* it usually doesn't work correctly on anything but sockets and pipes */ + flags &= ~EVBACKEND_KQUEUE; +#endif +#ifdef __APPLE__ + // flags &= ~EVBACKEND_KQUEUE; for documentation + flags &= ~EVBACKEND_POLL; +#endif + + return flags; +} + +unsigned int +ev_embeddable_backends (void) +{ + return EVBACKEND_EPOLL + | EVBACKEND_KQUEUE + | EVBACKEND_PORT; +} + +unsigned int +ev_backend (EV_P) +{ + return backend; +} + +unsigned int +ev_loop_count (EV_P) +{ + return loop_count; +} + +static void noinline loop_init (EV_P_ unsigned int flags) { - if (!method) + if (!backend) { #if EV_USE_MONOTONIC { @@ -773,24 +912,40 @@ loop_init (EV_P_ unsigned int flags) now_floor = mn_now; rtmn_diff = ev_rt_now - mn_now; - if (!(flags & EVFLAG_NOENV) && !enable_secure () && getenv ("LIBEV_FLAGS")) + /* pid check not overridable via env */ +#ifndef _WIN32 + if (flags & EVFLAG_FORKCHECK) + curpid = getpid (); +#endif + + if (!(flags & EVFLAG_NOENV) + && !enable_secure () + && getenv ("LIBEV_FLAGS")) flags = atoi (getenv ("LIBEV_FLAGS")); - if (!(flags & 0x0000ffff)) - flags |= 0x0000ffff; + if (!(flags & 0x0000ffffUL)) + flags |= ev_recommended_backends (); + + backend = 0; + backend_fd = -1; +#if EV_USE_INOTIFY + fs_fd = -2; +#endif - method = 0; +#if EV_USE_PORT + if (!backend && (flags & EVBACKEND_PORT )) backend = port_init (EV_A_ flags); +#endif #if EV_USE_KQUEUE - if (!method && (flags & EVMETHOD_KQUEUE)) method = kqueue_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_KQUEUE)) backend = kqueue_init (EV_A_ flags); #endif #if EV_USE_EPOLL - if (!method && (flags & EVMETHOD_EPOLL )) method = epoll_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_EPOLL )) backend = epoll_init (EV_A_ flags); #endif #if EV_USE_POLL - if (!method && (flags & EVMETHOD_POLL )) method = poll_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_POLL )) backend = poll_init (EV_A_ flags); #endif #if EV_USE_SELECT - if (!method && (flags & EVMETHOD_SELECT)) method = select_init (EV_A_ flags); + if (!backend && (flags & EVBACKEND_SELECT)) backend = select_init (EV_A_ flags); #endif ev_init (&sigev, sigcb); @@ -798,22 +953,33 @@ loop_init (EV_P_ unsigned int flags) } } -void +static void noinline loop_destroy (EV_P) { int i; +#if EV_USE_INOTIFY + if (fs_fd >= 0) + close (fs_fd); +#endif + + if (backend_fd >= 0) + close (backend_fd); + +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_destroy (EV_A); +#endif #if EV_USE_KQUEUE - if (method == EVMETHOD_KQUEUE) kqueue_destroy (EV_A); + if (backend == EVBACKEND_KQUEUE) kqueue_destroy (EV_A); #endif #if EV_USE_EPOLL - if (method == EVMETHOD_EPOLL ) epoll_destroy (EV_A); + if (backend == EVBACKEND_EPOLL ) epoll_destroy (EV_A); #endif #if EV_USE_POLL - if (method == EVMETHOD_POLL ) poll_destroy (EV_A); + if (backend == EVBACKEND_POLL ) poll_destroy (EV_A); #endif #if EV_USE_SELECT - if (method == EVMETHOD_SELECT) select_destroy (EV_A); + if (backend == EVBACKEND_SELECT) select_destroy (EV_A); #endif for (i = NUMPRI; i--; ) @@ -822,24 +988,32 @@ loop_destroy (EV_P) /* have to use the microsoft-never-gets-it-right macro */ array_free (fdchange, EMPTY0); array_free (timer, EMPTY0); -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE array_free (periodic, EMPTY0); #endif array_free (idle, EMPTY0); array_free (prepare, EMPTY0); array_free (check, EMPTY0); - method = 0; + backend = 0; } -static void +void inline_size infy_fork (EV_P); + +void inline_size loop_fork (EV_P) { -#if EV_USE_EPOLL - if (method == EVMETHOD_EPOLL ) epoll_fork (EV_A); +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_fork (EV_A); #endif #if EV_USE_KQUEUE - if (method == EVMETHOD_KQUEUE) kqueue_fork (EV_A); + if (backend == EVBACKEND_KQUEUE) kqueue_fork (EV_A); +#endif +#if EV_USE_EPOLL + if (backend == EVBACKEND_EPOLL ) epoll_fork (EV_A); +#endif +#if EV_USE_INOTIFY + infy_fork (EV_A); #endif if (ev_is_active (&sigev)) @@ -870,7 +1044,7 @@ ev_loop_new (unsigned int flags) loop_init (EV_A_ flags); - if (ev_method (EV_A)) + if (ev_backend (EV_A)) return loop; return 0; @@ -893,26 +1067,27 @@ ev_loop_fork (EV_P) #if EV_MULTIPLICITY struct ev_loop * +ev_default_loop_init (unsigned int flags) #else int -#endif ev_default_loop (unsigned int flags) +#endif { if (sigpipe [0] == sigpipe [1]) if (pipe (sigpipe)) return 0; - if (!default_loop) + if (!ev_default_loop_ptr) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop = &default_loop_struct; + struct ev_loop *loop = ev_default_loop_ptr = &default_loop_struct; #else - default_loop = 1; + ev_default_loop_ptr = 1; #endif loop_init (EV_A_ flags); - if (ev_method (EV_A)) + if (ev_backend (EV_A)) { siginit (EV_A); @@ -924,17 +1099,17 @@ ev_default_loop (unsigned int flags) #endif } else - default_loop = 0; + ev_default_loop_ptr = 0; } - return default_loop; + return ev_default_loop_ptr; } void ev_default_destroy (void) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop; + struct ev_loop *loop = ev_default_loop_ptr; #endif #ifndef _WIN32 @@ -955,16 +1130,16 @@ void ev_default_fork (void) { #if EV_MULTIPLICITY - struct ev_loop *loop = default_loop; + struct ev_loop *loop = ev_default_loop_ptr; #endif - if (method) + if (backend) postfork = 1; } /*****************************************************************************/ -static int +int inline_size any_pending (EV_P) { int pri; @@ -976,7 +1151,7 @@ any_pending (EV_P) return 0; } -static void +void inline_speed call_pending (EV_P) { int pri; @@ -986,22 +1161,24 @@ call_pending (EV_P) { ANPENDING *p = pendings [pri] + --pendingcnt [pri]; - if (p->w) + if (expect_true (p->w)) { + /*assert (("non-pending watcher on pending list", p->w->pending));*/ + p->w->pending = 0; EV_CB_INVOKE (p->w, p->events); } } } -static void +void inline_size timers_reify (EV_P) { while (timercnt && ((WT)timers [0])->at <= mn_now) { - struct ev_timer *w = timers [0]; + ev_timer *w = timers [0]; - assert (("inactive timer on timer heap detected", ev_is_active (w))); + /*assert (("inactive timer on timer heap detected", ev_is_active (w)));*/ /* first reschedule or stop timer */ if (w->repeat) @@ -1021,15 +1198,15 @@ timers_reify (EV_P) } } -#if EV_PERIODICS -static void +#if EV_PERIODIC_ENABLE +void inline_size periodics_reify (EV_P) { while (periodiccnt && ((WT)periodics [0])->at <= ev_rt_now) { - struct ev_periodic *w = periodics [0]; + ev_periodic *w = periodics [0]; - assert (("inactive timer on periodic heap detected", ev_is_active (w))); + /*assert (("inactive timer on periodic heap detected", ev_is_active (w)));*/ /* first reschedule or stop timer */ if (w->reschedule_cb) @@ -1051,7 +1228,7 @@ periodics_reify (EV_P) } } -static void +static void noinline periodics_reschedule (EV_P) { int i; @@ -1059,7 +1236,7 @@ periodics_reschedule (EV_P) /* adjust periodics after time jump */ for (i = 0; i < periodiccnt; ++i) { - struct ev_periodic *w = periodics [i]; + ev_periodic *w = periodics [i]; if (w->reschedule_cb) ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); @@ -1073,7 +1250,7 @@ periodics_reschedule (EV_P) } #endif -inline int +int inline_size time_update_monotonic (EV_P) { mn_now = get_clock (); @@ -1091,7 +1268,7 @@ time_update_monotonic (EV_P) } } -static void +void inline_size time_update (EV_P) { int i; @@ -1103,7 +1280,15 @@ time_update (EV_P) { ev_tstamp odiff = rtmn_diff; - for (i = 4; --i; ) /* loop a few times, before making important decisions */ + /* loop a few times, before making important decisions. + * on the choice of "4": one iteration isn't enough, + * in case we get preempted during the calls to + * ev_time and get_clock. a second call is almost guaranteed + * to succeed in that case, though. and looping a few more times + * doesn't hurt either as we only do this on time-jumps or + * in the unlikely event of having been preempted here. + */ + for (i = 4; --i; ) { rtmn_diff = ev_rt_now - mn_now; @@ -1115,7 +1300,7 @@ time_update (EV_P) now_floor = mn_now; } -# if EV_PERIODICS +# if EV_PERIODIC_ENABLE periodics_reschedule (EV_A); # endif /* no timer adjustment, as the monotonic clock doesn't jump */ @@ -1129,11 +1314,11 @@ time_update (EV_P) if (expect_false (mn_now > ev_rt_now || mn_now < ev_rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP)) { -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE periodics_reschedule (EV_A); #endif - /* adjust timers. this is easy, as the offset is the same for all */ + /* adjust timers. this is easy, as the offset is the same for all of them */ for (i = 0; i < timercnt; ++i) ((WT)timers [i])->at += ev_rt_now - mn_now; } @@ -1159,11 +1344,33 @@ static int loop_done; void ev_loop (EV_P_ int flags) { - double block; - loop_done = flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK) ? 1 : 0; + loop_done = flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK) + ? EVUNLOOP_ONE + : EVUNLOOP_CANCEL; + + call_pending (EV_A); /* in case we recurse, ensure ordering stays nice and clean */ do { +#ifndef _WIN32 + if (expect_false (curpid)) /* penalise the forking check even more */ + if (expect_false (getpid () != curpid)) + { + curpid = getpid (); + postfork = 1; + } +#endif + +#if EV_FORK_ENABLE + /* we might have forked, so queue fork handlers */ + if (expect_false (postfork)) + if (forkcnt) + { + queue_events (EV_A_ (W *)forks, forkcnt, EV_FORK); + call_pending (EV_A); + } +#endif + /* queue check watchers (and execute them) */ if (expect_false (preparecnt)) { @@ -1171,6 +1378,9 @@ ev_loop (EV_P_ int flags) call_pending (EV_A); } + if (expect_false (!activecnt)) + break; + /* we might have forked, so reify kernel state if necessary */ if (expect_false (postfork)) loop_fork (EV_A); @@ -1179,67 +1389,71 @@ ev_loop (EV_P_ int flags) fd_reify (EV_A); /* calculate blocking time */ + { + ev_tstamp block; - /* we only need this for !monotonic clock or timers, but as we basically - always have timers, we just calculate it always */ + if (expect_false (flags & EVLOOP_NONBLOCK || idlecnt || !activecnt)) + block = 0.; /* do not block at all */ + else + { + /* update time to cancel out callback processing overhead */ #if EV_USE_MONOTONIC - if (expect_true (have_monotonic)) - time_update_monotonic (EV_A); - else + if (expect_true (have_monotonic)) + time_update_monotonic (EV_A); + else #endif - { - ev_rt_now = ev_time (); - mn_now = ev_rt_now; - } + { + ev_rt_now = ev_time (); + mn_now = ev_rt_now; + } - if (flags & EVLOOP_NONBLOCK || idlecnt) - block = 0.; - else - { - block = MAX_BLOCKTIME; + block = MAX_BLOCKTIME; - if (timercnt) - { - ev_tstamp to = ((WT)timers [0])->at - mn_now + method_fudge; - if (block > to) block = to; - } + if (timercnt) + { + ev_tstamp to = ((WT)timers [0])->at - mn_now + backend_fudge; + if (block > to) block = to; + } -#if EV_PERIODICS - if (periodiccnt) - { - ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + method_fudge; - if (block > to) block = to; - } +#if EV_PERIODIC_ENABLE + if (periodiccnt) + { + ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + backend_fudge; + if (block > to) block = to; + } #endif - if (block < 0.) block = 0.; - } + if (expect_false (block < 0.)) block = 0.; + } - method_poll (EV_A_ block); + ++loop_count; + backend_poll (EV_A_ block); + } /* update ev_rt_now, do magic */ time_update (EV_A); /* queue pending timers and reschedule them */ timers_reify (EV_A); /* relative timers called last */ -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE periodics_reify (EV_A); /* absolute timers called first */ #endif - /* queue idle watchers unless io or timers are pending */ + /* queue idle watchers unless other events are pending */ if (idlecnt && !any_pending (EV_A)) queue_events (EV_A_ (W *)idles, idlecnt, EV_IDLE); /* queue check watchers, to be executed first */ - if (checkcnt) + if (expect_false (checkcnt)) queue_events (EV_A_ (W *)checks, checkcnt, EV_CHECK); call_pending (EV_A); + } - while (activecnt && !loop_done); + while (expect_true (activecnt && !loop_done)); - if (loop_done != 2) - loop_done = 0; + if (loop_done == EVUNLOOP_ONE) + loop_done = EVUNLOOP_CANCEL; } void @@ -1250,14 +1464,14 @@ ev_unloop (EV_P_ int how) /*****************************************************************************/ -inline void +void inline_size wlist_add (WL *head, WL elem) { elem->next = *head; *head = elem; } -inline void +void inline_size wlist_del (WL *head, WL elem) { while (*head) @@ -1272,7 +1486,7 @@ wlist_del (WL *head, WL elem) } } -inline void +void inline_speed ev_clear_pending (EV_P_ W w) { if (w->pending) @@ -1282,7 +1496,7 @@ ev_clear_pending (EV_P_ W w) } } -inline void +void inline_speed ev_start (EV_P_ W w, int active) { if (w->priority < EV_MINPRI) w->priority = EV_MINPRI; @@ -1292,7 +1506,7 @@ ev_start (EV_P_ W w, int active) ev_ref (EV_A); } -inline void +void inline_size ev_stop (EV_P_ W w) { ev_unref (EV_A); @@ -1302,11 +1516,11 @@ ev_stop (EV_P_ W w) /*****************************************************************************/ void -ev_io_start (EV_P_ struct ev_io *w) +ev_io_start (EV_P_ ev_io *w) { int fd = w->fd; - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; assert (("ev_io_start called with negative fd", fd >= 0)); @@ -1319,10 +1533,10 @@ ev_io_start (EV_P_ struct ev_io *w) } void -ev_io_stop (EV_P_ struct ev_io *w) +ev_io_stop (EV_P_ ev_io *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; assert (("ev_io_start called with illegal fd (must stay constant after start!)", w->fd >= 0 && w->fd < anfdmax)); @@ -1334,9 +1548,9 @@ ev_io_stop (EV_P_ struct ev_io *w) } void -ev_timer_start (EV_P_ struct ev_timer *w) +ev_timer_start (EV_P_ ev_timer *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; ((WT)w)->at += mn_now; @@ -1344,27 +1558,31 @@ ev_timer_start (EV_P_ struct ev_timer *w) assert (("ev_timer_start called with negative timer repeat value", w->repeat >= 0.)); ev_start (EV_A_ (W)w, ++timercnt); - array_needsize (struct ev_timer *, timers, timermax, timercnt, EMPTY2); + array_needsize (ev_timer *, timers, timermax, timercnt, EMPTY2); timers [timercnt - 1] = w; upheap ((WT *)timers, timercnt - 1); - assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w)); + /*assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w));*/ } void -ev_timer_stop (EV_P_ struct ev_timer *w) +ev_timer_stop (EV_P_ ev_timer *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w)); - if (((W)w)->active < timercnt--) - { - timers [((W)w)->active - 1] = timers [timercnt]; - adjustheap ((WT *)timers, timercnt, ((W)w)->active - 1); - } + { + int active = ((W)w)->active; + + if (expect_true (--active < --timercnt)) + { + timers [active] = timers [timercnt]; + adjustheap ((WT *)timers, timercnt, active); + } + } ((WT)w)->at -= mn_now; @@ -1372,7 +1590,7 @@ ev_timer_stop (EV_P_ struct ev_timer *w) } void -ev_timer_again (EV_P_ struct ev_timer *w) +ev_timer_again (EV_P_ ev_timer *w) { if (ev_is_active (w)) { @@ -1391,11 +1609,11 @@ ev_timer_again (EV_P_ struct ev_timer *w) } } -#if EV_PERIODICS +#if EV_PERIODIC_ENABLE void -ev_periodic_start (EV_P_ struct ev_periodic *w) +ev_periodic_start (EV_P_ ev_periodic *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; if (w->reschedule_cb) @@ -1408,33 +1626,37 @@ ev_periodic_start (EV_P_ struct ev_periodic *w) } ev_start (EV_A_ (W)w, ++periodiccnt); - array_needsize (struct ev_periodic *, periodics, periodicmax, periodiccnt, EMPTY2); + array_needsize (ev_periodic *, periodics, periodicmax, periodiccnt, EMPTY2); periodics [periodiccnt - 1] = w; upheap ((WT *)periodics, periodiccnt - 1); - assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w)); + /*assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w));*/ } void -ev_periodic_stop (EV_P_ struct ev_periodic *w) +ev_periodic_stop (EV_P_ ev_periodic *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w)); - if (((W)w)->active < periodiccnt--) - { - periodics [((W)w)->active - 1] = periodics [periodiccnt]; - adjustheap ((WT *)periodics, periodiccnt, ((W)w)->active - 1); - } + { + int active = ((W)w)->active; + + if (expect_true (--active < --periodiccnt)) + { + periodics [active] = periodics [periodiccnt]; + adjustheap ((WT *)periodics, periodiccnt, active); + } + } ev_stop (EV_A_ (W)w); } void -ev_periodic_again (EV_P_ struct ev_periodic *w) +ev_periodic_again (EV_P_ ev_periodic *w) { /* TODO: use adjustheap and recalculation */ ev_periodic_stop (EV_A_ w); @@ -1442,149 +1664,493 @@ ev_periodic_again (EV_P_ struct ev_periodic *w) } #endif +#ifndef SA_RESTART +# define SA_RESTART 0 +#endif + void -ev_idle_start (EV_P_ struct ev_idle *w) +ev_signal_start (EV_P_ ev_signal *w) { - if (ev_is_active (w)) +#if EV_MULTIPLICITY + assert (("signal watchers are only supported in the default loop", loop == ev_default_loop_ptr)); +#endif + if (expect_false (ev_is_active (w))) + return; + + assert (("ev_signal_start called with illegal signal number", w->signum > 0)); + + ev_start (EV_A_ (W)w, 1); + array_needsize (ANSIG, signals, signalmax, w->signum, signals_init); + wlist_add ((WL *)&signals [w->signum - 1].head, (WL)w); + + if (!((WL)w)->next) + { +#if _WIN32 + signal (w->signum, sighandler); +#else + struct sigaction sa; + sa.sa_handler = sighandler; + sigfillset (&sa.sa_mask); + sa.sa_flags = SA_RESTART; /* if restarting works we save one iteration */ + sigaction (w->signum, &sa, 0); +#endif + } +} + +void +ev_signal_stop (EV_P_ ev_signal *w) +{ + ev_clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + wlist_del ((WL *)&signals [w->signum - 1].head, (WL)w); + ev_stop (EV_A_ (W)w); + + if (!signals [w->signum - 1].head) + signal (w->signum, SIG_DFL); +} + +void +ev_child_start (EV_P_ ev_child *w) +{ +#if EV_MULTIPLICITY + assert (("child watchers are only supported in the default loop", loop == ev_default_loop_ptr)); +#endif + if (expect_false (ev_is_active (w))) + return; + + ev_start (EV_A_ (W)w, 1); + wlist_add ((WL *)&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); +} + +void +ev_child_stop (EV_P_ ev_child *w) +{ + ev_clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + wlist_del ((WL *)&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); + ev_stop (EV_A_ (W)w); +} + +#if EV_STAT_ENABLE + +# ifdef _WIN32 +# undef lstat +# define lstat(a,b) _stati64 (a,b) +# endif + +#define DEF_STAT_INTERVAL 5.0074891 +#define MIN_STAT_INTERVAL 0.1074891 + +static void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents); + +#if EV_USE_INOTIFY +# define EV_INOTIFY_BUFSIZE 8192 + +static void noinline +infy_add (EV_P_ ev_stat *w) +{ + w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); + + if (w->wd < 0) + { + ev_timer_start (EV_A_ &w->timer); /* this is not race-free, so we still need to recheck periodically */ + + /* monitor some parent directory for speedup hints */ + if ((errno == ENOENT || errno == EACCES) && strlen (w->path) < 4096) + { + char path [4096]; + strcpy (path, w->path); + + do + { + int mask = IN_MASK_ADD | IN_DELETE_SELF | IN_MOVE_SELF + | (errno == EACCES ? IN_ATTRIB : IN_CREATE | IN_MOVED_TO); + + char *pend = strrchr (path, '/'); + + if (!pend) + break; /* whoops, no '/', complain to your admin */ + + *pend = 0; + w->wd = inotify_add_watch (fs_fd, path, mask); + } + while (w->wd < 0 && (errno == ENOENT || errno == EACCES)); + } + } + else + ev_timer_stop (EV_A_ &w->timer); /* we can watch this in a race-free way */ + + if (w->wd >= 0) + wlist_add (&fs_hash [w->wd & (EV_INOTIFY_HASHSIZE - 1)].head, (WL)w); +} + +static void noinline +infy_del (EV_P_ ev_stat *w) +{ + int slot; + int wd = w->wd; + + if (wd < 0) + return; + + w->wd = -2; + slot = wd & (EV_INOTIFY_HASHSIZE - 1); + wlist_del (&fs_hash [slot].head, (WL)w); + + /* remove this watcher, if others are watching it, they will rearm */ + inotify_rm_watch (fs_fd, wd); +} + +static void noinline +infy_wd (EV_P_ int slot, int wd, struct inotify_event *ev) +{ + if (slot < 0) + /* overflow, need to check for all hahs slots */ + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + infy_wd (EV_A_ slot, wd, ev); + else + { + WL w_; + + for (w_ = fs_hash [slot & (EV_INOTIFY_HASHSIZE - 1)].head; w_; ) + { + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us remove this watcher and all before it */ + + if (w->wd == wd || wd == -1) + { + if (ev->mask & (IN_IGNORED | IN_UNMOUNT | IN_DELETE_SELF)) + { + w->wd = -1; + infy_add (EV_A_ w); /* re-add, no matter what */ + } + + stat_timer_cb (EV_A_ &w->timer, 0); + } + } + } +} + +static void +infy_cb (EV_P_ ev_io *w, int revents) +{ + char buf [EV_INOTIFY_BUFSIZE]; + struct inotify_event *ev = (struct inotify_event *)buf; + int ofs; + int len = read (fs_fd, buf, sizeof (buf)); + + for (ofs = 0; ofs < len; ofs += sizeof (struct inotify_event) + ev->len) + infy_wd (EV_A_ ev->wd, ev->wd, ev); +} + +void inline_size +infy_init (EV_P) +{ + if (fs_fd != -2) + return; + + fs_fd = inotify_init (); + + if (fs_fd >= 0) + { + ev_io_init (&fs_w, infy_cb, fs_fd, EV_READ); + ev_set_priority (&fs_w, EV_MAXPRI); + ev_io_start (EV_A_ &fs_w); + } +} + +void inline_size +infy_fork (EV_P) +{ + int slot; + + if (fs_fd < 0) + return; + + close (fs_fd); + fs_fd = inotify_init (); + + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + { + WL w_ = fs_hash [slot].head; + fs_hash [slot].head = 0; + + while (w_) + { + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us add this watcher */ + + w->wd = -1; + + if (fs_fd >= 0) + infy_add (EV_A_ w); /* re-add, no matter what */ + else + ev_timer_start (EV_A_ &w->timer); + } + + } +} + +#endif + +void +ev_stat_stat (EV_P_ ev_stat *w) +{ + if (lstat (w->path, &w->attr) < 0) + w->attr.st_nlink = 0; + else if (!w->attr.st_nlink) + w->attr.st_nlink = 1; +} + +static void noinline +stat_timer_cb (EV_P_ ev_timer *w_, int revents) +{ + ev_stat *w = (ev_stat *)(((char *)w_) - offsetof (ev_stat, timer)); + + /* we copy this here each the time so that */ + /* prev has the old value when the callback gets invoked */ + w->prev = w->attr; + ev_stat_stat (EV_A_ w); + + /* memcmp doesn't work on netbsd, they.... do stuff to their struct stat */ + if ( + w->prev.st_dev != w->attr.st_dev + || w->prev.st_ino != w->attr.st_ino + || w->prev.st_mode != w->attr.st_mode + || w->prev.st_nlink != w->attr.st_nlink + || w->prev.st_uid != w->attr.st_uid + || w->prev.st_gid != w->attr.st_gid + || w->prev.st_rdev != w->attr.st_rdev + || w->prev.st_size != w->attr.st_size + || w->prev.st_atime != w->attr.st_atime + || w->prev.st_mtime != w->attr.st_mtime + || w->prev.st_ctime != w->attr.st_ctime + ) { + #if EV_USE_INOTIFY + infy_del (EV_A_ w); + infy_add (EV_A_ w); + ev_stat_stat (EV_A_ w); /* avoid race... */ + #endif + + ev_feed_event (EV_A_ w, EV_STAT); + } +} + +void +ev_stat_start (EV_P_ ev_stat *w) +{ + if (expect_false (ev_is_active (w))) + return; + + /* since we use memcmp, we need to clear any padding data etc. */ + memset (&w->prev, 0, sizeof (ev_statdata)); + memset (&w->attr, 0, sizeof (ev_statdata)); + + ev_stat_stat (EV_A_ w); + + if (w->interval < MIN_STAT_INTERVAL) + w->interval = w->interval ? MIN_STAT_INTERVAL : DEF_STAT_INTERVAL; + + ev_timer_init (&w->timer, stat_timer_cb, w->interval, w->interval); + ev_set_priority (&w->timer, ev_priority (w)); + +#if EV_USE_INOTIFY + infy_init (EV_A); + + if (fs_fd >= 0) + infy_add (EV_A_ w); + else +#endif + ev_timer_start (EV_A_ &w->timer); + + ev_start (EV_A_ (W)w, 1); +} + +void +ev_stat_stop (EV_P_ ev_stat *w) +{ + ev_clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + +#if EV_USE_INOTIFY + infy_del (EV_A_ w); +#endif + ev_timer_stop (EV_A_ &w->timer); + + ev_stop (EV_A_ (W)w); +} +#endif + +void +ev_idle_start (EV_P_ ev_idle *w) +{ + if (expect_false (ev_is_active (w))) return; ev_start (EV_A_ (W)w, ++idlecnt); - array_needsize (struct ev_idle *, idles, idlemax, idlecnt, EMPTY2); + array_needsize (ev_idle *, idles, idlemax, idlecnt, EMPTY2); idles [idlecnt - 1] = w; } void -ev_idle_stop (EV_P_ struct ev_idle *w) +ev_idle_stop (EV_P_ ev_idle *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; - idles [((W)w)->active - 1] = idles [--idlecnt]; + { + int active = ((W)w)->active; + idles [active - 1] = idles [--idlecnt]; + ((W)idles [active - 1])->active = active; + } + ev_stop (EV_A_ (W)w); } void -ev_prepare_start (EV_P_ struct ev_prepare *w) +ev_prepare_start (EV_P_ ev_prepare *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; ev_start (EV_A_ (W)w, ++preparecnt); - array_needsize (struct ev_prepare *, prepares, preparemax, preparecnt, EMPTY2); + array_needsize (ev_prepare *, prepares, preparemax, preparecnt, EMPTY2); prepares [preparecnt - 1] = w; } void -ev_prepare_stop (EV_P_ struct ev_prepare *w) +ev_prepare_stop (EV_P_ ev_prepare *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; - prepares [((W)w)->active - 1] = prepares [--preparecnt]; + { + int active = ((W)w)->active; + prepares [active - 1] = prepares [--preparecnt]; + ((W)prepares [active - 1])->active = active; + } + ev_stop (EV_A_ (W)w); } void -ev_check_start (EV_P_ struct ev_check *w) +ev_check_start (EV_P_ ev_check *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; ev_start (EV_A_ (W)w, ++checkcnt); - array_needsize (struct ev_check *, checks, checkmax, checkcnt, EMPTY2); + array_needsize (ev_check *, checks, checkmax, checkcnt, EMPTY2); checks [checkcnt - 1] = w; } void -ev_check_stop (EV_P_ struct ev_check *w) +ev_check_stop (EV_P_ ev_check *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; - checks [((W)w)->active - 1] = checks [--checkcnt]; + { + int active = ((W)w)->active; + checks [active - 1] = checks [--checkcnt]; + ((W)checks [active - 1])->active = active; + } + ev_stop (EV_A_ (W)w); } -#ifndef SA_RESTART -# define SA_RESTART 0 -#endif +#if EV_EMBED_ENABLE +void noinline +ev_embed_sweep (EV_P_ ev_embed *w) +{ + ev_loop (w->loop, EVLOOP_NONBLOCK); +} + +static void +embed_cb (EV_P_ ev_io *io, int revents) +{ + ev_embed *w = (ev_embed *)(((char *)io) - offsetof (ev_embed, io)); + + if (ev_cb (w)) + ev_feed_event (EV_A_ (W)w, EV_EMBED); + else + ev_embed_sweep (loop, w); +} void -ev_signal_start (EV_P_ struct ev_signal *w) +ev_embed_start (EV_P_ ev_embed *w) { -#if EV_MULTIPLICITY - assert (("signal watchers are only supported in the default loop", loop == default_loop)); -#endif - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - assert (("ev_signal_start called with illegal signal number", w->signum > 0)); + { + struct ev_loop *loop = w->loop; + assert (("loop to be embedded is not embeddable", backend & ev_embeddable_backends ())); + ev_io_init (&w->io, embed_cb, backend_fd, EV_READ); + } - ev_start (EV_A_ (W)w, 1); - array_needsize (ANSIG, signals, signalmax, w->signum, signals_init); - wlist_add ((WL *)&signals [w->signum - 1].head, (WL)w); + ev_set_priority (&w->io, ev_priority (w)); + ev_io_start (EV_A_ &w->io); - if (!((WL)w)->next) - { -#if _WIN32 - signal (w->signum, sighandler); -#else - struct sigaction sa; - sa.sa_handler = sighandler; - sigfillset (&sa.sa_mask); - sa.sa_flags = SA_RESTART; /* if restarting works we save one iteration */ - sigaction (w->signum, &sa, 0); -#endif - } + ev_start (EV_A_ (W)w, 1); } void -ev_signal_stop (EV_P_ struct ev_signal *w) +ev_embed_stop (EV_P_ ev_embed *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; - wlist_del ((WL *)&signals [w->signum - 1].head, (WL)w); - ev_stop (EV_A_ (W)w); + ev_io_stop (EV_A_ &w->io); - if (!signals [w->signum - 1].head) - signal (w->signum, SIG_DFL); + ev_stop (EV_A_ (W)w); } +#endif +#if EV_FORK_ENABLE void -ev_child_start (EV_P_ struct ev_child *w) +ev_fork_start (EV_P_ ev_fork *w) { -#if EV_MULTIPLICITY - assert (("child watchers are only supported in the default loop", loop == default_loop)); -#endif - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - ev_start (EV_A_ (W)w, 1); - wlist_add ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); + ev_start (EV_A_ (W)w, ++forkcnt); + array_needsize (ev_fork *, forks, forkmax, forkcnt, EMPTY2); + forks [forkcnt - 1] = w; } void -ev_child_stop (EV_P_ struct ev_child *w) +ev_fork_stop (EV_P_ ev_fork *w) { ev_clear_pending (EV_A_ (W)w); - if (!ev_is_active (w)) + if (expect_false (!ev_is_active (w))) return; - wlist_del ((WL *)&childs [w->pid & (PID_HASHSIZE - 1)], (WL)w); + { + int active = ((W)w)->active; + forks [active - 1] = forks [--forkcnt]; + ((W)forks [active - 1])->active = active; + } + ev_stop (EV_A_ (W)w); } +#endif /*****************************************************************************/ struct ev_once { - struct ev_io io; - struct ev_timer to; + ev_io io; + ev_timer to; void (*cb)(int revents, void *arg); void *arg; }; @@ -1603,13 +2169,13 @@ once_cb (EV_P_ struct ev_once *once, int revents) } static void -once_cb_io (EV_P_ struct ev_io *w, int revents) +once_cb_io (EV_P_ ev_io *w, int revents) { once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, io)), revents); } static void -once_cb_to (EV_P_ struct ev_timer *w, int revents) +once_cb_to (EV_P_ ev_timer *w, int revents) { once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, to)), revents); } @@ -1619,26 +2185,27 @@ ev_once (EV_P_ int fd, int events, ev_tstamp timeout, void (*cb)(int revents, vo { struct ev_once *once = (struct ev_once *)ev_malloc (sizeof (struct ev_once)); - if (!once) - cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); - else + if (expect_false (!once)) { - once->cb = cb; - once->arg = arg; + cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); + return; + } - ev_init (&once->io, once_cb_io); - if (fd >= 0) - { - ev_io_set (&once->io, fd, events); - ev_io_start (EV_A_ &once->io); - } + once->cb = cb; + once->arg = arg; - ev_init (&once->to, once_cb_to); - if (timeout >= 0.) - { - ev_timer_set (&once->to, timeout, 0.); - ev_timer_start (EV_A_ &once->to); - } + ev_init (&once->io, once_cb_io); + if (fd >= 0) + { + ev_io_set (&once->io, fd, events); + ev_io_start (EV_A_ &once->io); + } + + ev_init (&once->to, once_cb_to); + if (timeout >= 0.) + { + ev_timer_set (&once->to, timeout, 0.); + ev_timer_start (EV_A_ &once->to); } }