X-Git-Url: https://git.llucax.com/software/libev.git/blobdiff_plain/04c8cb9913a370451296919b6dd2567793de9a75..f279f2b7372a6523b4055dea41bc75e0a3d7b253:/ev.c diff --git a/ev.c b/ev.c index 1e54f21..d024017 100644 --- a/ev.c +++ b/ev.c @@ -1,73 +1,412 @@ +/* + * libev event processing core, watcher management + * + * Copyright (c) 2007 Marc Alexander Lehmann + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without modifica- + * tion, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, + * this list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR IMPLIED + * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MER- + * CHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO + * EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPE- + * CIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, + * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; + * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTH- + * ERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED + * OF THE POSSIBILITY OF SUCH DAMAGE. + * + * Alternatively, the contents of this file may be used under the terms of + * the GNU General Public License ("GPL") version 2 or any later version, + * in which case the provisions of the GPL are applicable instead of + * the above. If you wish to allow the use of your version of this file + * only under the terms of the GPL and not to allow others to use your + * version of this file under the BSD license, indicate your decision + * by deleting the provisions above and replace them with the notice + * and other provisions required by the GPL. If you do not delete the + * provisions above, a recipient may use your version of this file under + * either the BSD or the GPL. + */ + +#ifdef __cplusplus +extern "C" { +#endif + +#ifndef EV_STANDALONE +# ifdef EV_CONFIG_H +# include EV_CONFIG_H +# else +# include "config.h" +# endif + +# if HAVE_CLOCK_GETTIME +# ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 1 +# endif +# ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 1 +# endif +# else +# ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 0 +# endif +# ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 +# endif +# endif + +# ifndef EV_USE_NANOSLEEP +# if HAVE_NANOSLEEP +# define EV_USE_NANOSLEEP 1 +# else +# define EV_USE_NANOSLEEP 0 +# endif +# endif + +# ifndef EV_USE_SELECT +# if HAVE_SELECT && HAVE_SYS_SELECT_H +# define EV_USE_SELECT 1 +# else +# define EV_USE_SELECT 0 +# endif +# endif + +# ifndef EV_USE_POLL +# if HAVE_POLL && HAVE_POLL_H +# define EV_USE_POLL 1 +# else +# define EV_USE_POLL 0 +# endif +# endif + +# ifndef EV_USE_EPOLL +# if HAVE_EPOLL_CTL && HAVE_SYS_EPOLL_H +# define EV_USE_EPOLL 1 +# else +# define EV_USE_EPOLL 0 +# endif +# endif + +# ifndef EV_USE_KQUEUE +# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H +# define EV_USE_KQUEUE 1 +# else +# define EV_USE_KQUEUE 0 +# endif +# endif + +# ifndef EV_USE_PORT +# if HAVE_PORT_H && HAVE_PORT_CREATE +# define EV_USE_PORT 1 +# else +# define EV_USE_PORT 0 +# endif +# endif + +# ifndef EV_USE_INOTIFY +# if HAVE_INOTIFY_INIT && HAVE_SYS_INOTIFY_H +# define EV_USE_INOTIFY 1 +# else +# define EV_USE_INOTIFY 0 +# endif +# endif + +#endif + #include #include -#include #include -#include +#include #include #include #include -#include +#include #include -#define HAVE_EPOLL 1 +#include + +#ifdef EV_H +# include EV_H +#else +# include "ev.h" +#endif + +#ifndef _WIN32 +# include +# include +# include +#else +# define WIN32_LEAN_AND_MEAN +# include +# ifndef EV_SELECT_IS_WINSOCKET +# define EV_SELECT_IS_WINSOCKET 1 +# endif +#endif + +/**/ + +#ifndef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 0 +#endif + +#ifndef EV_USE_REALTIME +# define EV_USE_REALTIME 0 +#endif + +#ifndef EV_USE_NANOSLEEP +# define EV_USE_NANOSLEEP 0 +#endif + +#ifndef EV_USE_SELECT +# define EV_USE_SELECT 1 +#endif + +#ifndef EV_USE_POLL +# ifdef _WIN32 +# define EV_USE_POLL 0 +# else +# define EV_USE_POLL 1 +# endif +#endif + +#ifndef EV_USE_EPOLL +# define EV_USE_EPOLL 0 +#endif + +#ifndef EV_USE_KQUEUE +# define EV_USE_KQUEUE 0 +#endif + +#ifndef EV_USE_PORT +# define EV_USE_PORT 0 +#endif + +#ifndef EV_USE_INOTIFY +# define EV_USE_INOTIFY 0 +#endif -#ifndef HAVE_MONOTONIC -# ifdef CLOCK_MONOTONIC -# define HAVE_MONOTONIC 1 +#ifndef EV_PID_HASHSIZE +# if EV_MINIMAL +# define EV_PID_HASHSIZE 1 +# else +# define EV_PID_HASHSIZE 16 # endif #endif -#ifndef HAVE_SELECT -# define HAVE_SELECT 1 +#ifndef EV_INOTIFY_HASHSIZE +# if EV_MINIMAL +# define EV_INOTIFY_HASHSIZE 1 +# else +# define EV_INOTIFY_HASHSIZE 16 +# endif +#endif + +/**/ + +#ifndef CLOCK_MONOTONIC +# undef EV_USE_MONOTONIC +# define EV_USE_MONOTONIC 0 +#endif + +#ifndef CLOCK_REALTIME +# undef EV_USE_REALTIME +# define EV_USE_REALTIME 0 +#endif + +#if !EV_STAT_ENABLE +# undef EV_USE_INOTIFY +# define EV_USE_INOTIFY 0 #endif -#ifndef HAVE_EPOLL -# define HAVE_EPOLL 0 +#if !EV_USE_NANOSLEEP +# ifndef _WIN32 +# include +# endif +#endif + +#if EV_USE_INOTIFY +# include #endif -#ifndef HAVE_REALTIME -# define HAVE_REALTIME 1 /* posix requirement, but might be slower */ +#if EV_SELECT_IS_WINSOCKET +# include #endif +/**/ + +/* + * This is used to avoid floating point rounding problems. + * It is added to ev_rt_now when scheduling periodics + * to ensure progress, time-wise, even when rounding + * errors are against us. + * This value is good at least till the year 4000. + * Better solutions welcome. + */ +#define TIME_EPSILON 0.0001220703125 /* 1/8192 */ + #define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ -#define MAX_BLOCKTIME 60. +#define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ +/*#define CLEANUP_INTERVAL (MAX_BLOCKTIME * 5.) /* how often to try to free memory and re-check fds, TODO */ + +#if __GNUC__ >= 4 +# define expect(expr,value) __builtin_expect ((expr),(value)) +# define noinline __attribute__ ((noinline)) +#else +# define expect(expr,value) (expr) +# define noinline +# if __STDC_VERSION__ < 199901L +# define inline +# endif +#endif -#include "ev.h" +#define expect_false(expr) expect ((expr) != 0, 0) +#define expect_true(expr) expect ((expr) != 0, 1) +#define inline_size static inline -struct ev_watcher { - EV_WATCHER (ev_watcher); -}; +#if EV_MINIMAL +# define inline_speed static noinline +#else +# define inline_speed static inline +#endif -struct ev_watcher_list { - EV_WATCHER_LIST (ev_watcher_list); -}; +#define NUMPRI (EV_MAXPRI - EV_MINPRI + 1) +#define ABSPRI(w) (((W)w)->priority - EV_MINPRI) -struct ev_watcher_time { - EV_WATCHER_TIME (ev_watcher_time); -}; +#define EMPTY /* required for microsofts broken pseudo-c compiler */ +#define EMPTY2(a,b) /* used to suppress some warnings */ + +typedef ev_watcher *W; +typedef ev_watcher_list *WL; +typedef ev_watcher_time *WT; + +#if EV_USE_MONOTONIC +/* sig_atomic_t is used to avoid per-thread variables or locking but still */ +/* giving it a reasonably high chance of working on typical architetcures */ +static sig_atomic_t have_monotonic; /* did clock_gettime (CLOCK_MONOTONIC) work? */ +#endif + +#ifdef _WIN32 +# include "ev_win32.c" +#endif + +/*****************************************************************************/ + +static void (*syserr_cb)(const char *msg); + +void +ev_set_syserr_cb (void (*cb)(const char *msg)) +{ + syserr_cb = cb; +} + +static void noinline +syserr (const char *msg) +{ + if (!msg) + msg = "(libev) system error"; + + if (syserr_cb) + syserr_cb (msg); + else + { + perror (msg); + abort (); + } +} + +static void *(*alloc)(void *ptr, long size); + +void +ev_set_allocator (void *(*cb)(void *ptr, long size)) +{ + alloc = cb; +} + +inline_speed void * +ev_realloc (void *ptr, long size) +{ + ptr = alloc ? alloc (ptr, size) : realloc (ptr, size); -typedef struct ev_watcher *W; -typedef struct ev_watcher_list *WL; -typedef struct ev_watcher_time *WT; + if (!ptr && size) + { + fprintf (stderr, "libev: cannot allocate %ld bytes, aborting.", size); + abort (); + } + + return ptr; +} + +#define ev_malloc(size) ev_realloc (0, (size)) +#define ev_free(ptr) ev_realloc ((ptr), 0) + +/*****************************************************************************/ + +typedef struct +{ + WL head; + unsigned char events; + unsigned char reify; +#if EV_SELECT_IS_WINSOCKET + SOCKET handle; +#endif +} ANFD; + +typedef struct +{ + W w; + int events; +} ANPENDING; + +#if EV_USE_INOTIFY +typedef struct +{ + WL head; +} ANFS; +#endif + +#if EV_MULTIPLICITY + + struct ev_loop + { + ev_tstamp ev_rt_now; + #define ev_rt_now ((loop)->ev_rt_now) + #define VAR(name,decl) decl; + #include "ev_vars.h" + #undef VAR + }; + #include "ev_wrap.h" + + static struct ev_loop default_loop_struct; + struct ev_loop *ev_default_loop_ptr; + +#else -static ev_tstamp now, diff; /* monotonic clock */ -ev_tstamp ev_now; -int ev_method; + ev_tstamp ev_rt_now; + #define VAR(name,decl) static decl; + #include "ev_vars.h" + #undef VAR -static int have_monotonic; /* runtime */ + static int ev_default_loop_ptr; -static ev_tstamp method_fudge; /* stupid epoll-returns-early bug */ -static void (*method_modify)(int fd, int oev, int nev); -static void (*method_poll)(ev_tstamp timeout); +#endif /*****************************************************************************/ ev_tstamp ev_time (void) { -#if HAVE_REALTIME +#if EV_USE_REALTIME struct timespec ts; clock_gettime (CLOCK_REALTIME, &ts); return ts.tv_sec + ts.tv_nsec * 1e-9; @@ -78,11 +417,11 @@ ev_time (void) #endif } -static ev_tstamp +ev_tstamp inline_size get_clock (void) { -#if HAVE_MONOTONIC - if (have_monotonic) +#if EV_USE_MONOTONIC + if (expect_true (have_monotonic)) { struct timespec ts; clock_gettime (CLOCK_MONOTONIC, &ts); @@ -93,154 +432,357 @@ get_clock (void) return ev_time (); } -#define array_needsize(base,cur,cnt,init) \ - if ((cnt) > cur) \ - { \ - int newcnt = cur ? cur << 1 : 16; \ - fprintf (stderr, "resize(" # base ") from %d to %d\n", cur, newcnt);\ - base = realloc (base, sizeof (*base) * (newcnt)); \ - init (base + cur, newcnt - cur); \ - cur = newcnt; \ +#if EV_MULTIPLICITY +ev_tstamp +ev_now (EV_P) +{ + return ev_rt_now; +} +#endif + +void +ev_sleep (ev_tstamp delay) +{ + if (delay > 0.) + { +#if EV_USE_NANOSLEEP + struct timespec ts; + + ts.tv_sec = (time_t)delay; + ts.tv_nsec = (long)((delay - (ev_tstamp)(ts.tv_sec)) * 1e9); + + nanosleep (&ts, 0); +#elif defined(_WIN32) + Sleep (delay * 1e3); +#else + struct timeval tv; + + tv.tv_sec = (time_t)delay; + tv.tv_usec = (long)((delay - (ev_tstamp)(tv.tv_sec)) * 1e6); + + select (0, 0, 0, 0, &tv); +#endif } +} /*****************************************************************************/ -typedef struct +int inline_size +array_nextsize (int elem, int cur, int cnt) { - struct ev_io *head; - unsigned char wev, rev; /* want, received event set */ -} ANFD; + int ncur = cur + 1; + + do + ncur <<= 1; + while (cnt > ncur); -static ANFD *anfds; -static int anfdmax; + /* if size > 4096, round to 4096 - 4 * longs to accomodate malloc overhead */ + if (elem * ncur > 4096) + { + ncur *= elem; + ncur = (ncur + elem + 4095 + sizeof (void *) * 4) & ~4095; + ncur = ncur - sizeof (void *) * 4; + ncur /= elem; + } -static int *fdchanges; -static int fdchangemax, fdchangecnt; + return ncur; +} -static void -anfds_init (ANFD *base, int count) +static noinline void * +array_realloc (int elem, void *base, int *cur, int cnt) { - while (count--) + *cur = array_nextsize (elem, *cur, cnt); + return ev_realloc (base, elem * *cur); +} + +#define array_needsize(type,base,cur,cnt,init) \ + if (expect_false ((cnt) > (cur))) \ + { \ + int ocur_ = (cur); \ + (base) = (type *)array_realloc \ + (sizeof (type), (base), &(cur), (cnt)); \ + init ((base) + (ocur_), (cur) - ocur_); \ + } + +#if 0 +#define array_slim(type,stem) \ + if (stem ## max < array_roundsize (stem ## cnt >> 2)) \ + { \ + stem ## max = array_roundsize (stem ## cnt >> 1); \ + base = (type *)ev_realloc (base, sizeof (type) * (stem ## max));\ + fprintf (stderr, "slimmed down " # stem " to %d\n", stem ## max);/*D*/\ + } +#endif + +#define array_free(stem, idx) \ + ev_free (stem ## s idx); stem ## cnt idx = stem ## max idx = 0; + +/*****************************************************************************/ + +void noinline +ev_feed_event (EV_P_ void *w, int revents) +{ + W w_ = (W)w; + int pri = ABSPRI (w_); + + if (expect_false (w_->pending)) + pendings [pri][w_->pending - 1].events |= revents; + else { - base->head = 0; - base->wev = base->rev = EV_NONE; - ++base; + w_->pending = ++pendingcnt [pri]; + array_needsize (ANPENDING, pendings [pri], pendingmax [pri], w_->pending, EMPTY2); + pendings [pri][w_->pending - 1].w = w_; + pendings [pri][w_->pending - 1].events = revents; } } -typedef struct +void inline_speed +queue_events (EV_P_ W *events, int eventcnt, int type) { - W w; - int events; -} ANPENDING; + int i; + + for (i = 0; i < eventcnt; ++i) + ev_feed_event (EV_A_ events [i], type); +} -static ANPENDING *pendings; -static int pendingmax, pendingcnt; +/*****************************************************************************/ -static void -event (W w, int events) +void inline_size +anfds_init (ANFD *base, int count) { - w->pending = ++pendingcnt; - array_needsize (pendings, pendingmax, pendingcnt, ); - pendings [pendingcnt - 1].w = w; - pendings [pendingcnt - 1].events = events; + while (count--) + { + base->head = 0; + base->events = EV_NONE; + base->reify = 0; + + ++base; + } } -static void -fd_event (int fd, int events) +void inline_speed +fd_event (EV_P_ int fd, int revents) { ANFD *anfd = anfds + fd; - struct ev_io *w; + ev_io *w; - for (w = anfd->head; w; w = w->next) + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) { - int ev = w->events & events; + int ev = w->events & revents; if (ev) - event ((W)w, ev); + ev_feed_event (EV_A_ (W)w, ev); } } -static void -queue_events (W *events, int eventcnt, int type) +void +ev_feed_fd_event (EV_P_ int fd, int revents) +{ + if (fd >= 0 && fd < anfdmax) + fd_event (EV_A_ fd, revents); +} + +void inline_size +fd_reify (EV_P) { int i; - for (i = 0; i < eventcnt; ++i) - event (events [i], type); -} + for (i = 0; i < fdchangecnt; ++i) + { + int fd = fdchanges [i]; + ANFD *anfd = anfds + fd; + ev_io *w; -/*****************************************************************************/ + unsigned char events = 0; -static struct ev_timer **timers; -static int timermax, timercnt; + for (w = (ev_io *)anfd->head; w; w = (ev_io *)((WL)w)->next) + events |= (unsigned char)w->events; -static struct ev_periodic **periodics; -static int periodicmax, periodiccnt; +#if EV_SELECT_IS_WINSOCKET + if (events) + { + unsigned long argp; + #ifdef EV_FD_TO_WIN32_HANDLE + anfd->handle = EV_FD_TO_WIN32_HANDLE (fd); + #else + anfd->handle = _get_osfhandle (fd); + #endif + assert (("libev only supports socket fds in this configuration", ioctlsocket (anfd->handle, FIONREAD, &argp) == 0)); + } +#endif -static void -upheap (WT *timers, int k) + { + unsigned char o_events = anfd->events; + unsigned char o_reify = anfd->reify; + + anfd->reify = 0; + anfd->events = events; + + if (o_events != events || o_reify & EV_IOFDSET) + backend_modify (EV_A_ fd, o_events, events); + } + } + + fdchangecnt = 0; +} + +void inline_size +fd_change (EV_P_ int fd, int flags) { - WT w = timers [k]; + unsigned char reify = anfds [fd].reify; + anfds [fd].reify |= flags; - while (k && timers [k >> 1]->at > w->at) + if (expect_true (!reify)) { - timers [k] = timers [k >> 1]; - timers [k]->active = k + 1; - k >>= 1; + ++fdchangecnt; + array_needsize (int, fdchanges, fdchangemax, fdchangecnt, EMPTY2); + fdchanges [fdchangecnt - 1] = fd; } +} - timers [k] = w; - timers [k]->active = k + 1; +void inline_speed +fd_kill (EV_P_ int fd) +{ + ev_io *w; + while ((w = (ev_io *)anfds [fd].head)) + { + ev_io_stop (EV_A_ w); + ev_feed_event (EV_A_ (W)w, EV_ERROR | EV_READ | EV_WRITE); + } } -static void -downheap (WT *timers, int N, int k) +int inline_size +fd_valid (int fd) +{ +#ifdef _WIN32 + return _get_osfhandle (fd) != -1; +#else + return fcntl (fd, F_GETFD) != -1; +#endif +} + +/* called on EBADF to verify fds */ +static void noinline +fd_ebadf (EV_P) +{ + int fd; + + for (fd = 0; fd < anfdmax; ++fd) + if (anfds [fd].events) + if (!fd_valid (fd) == -1 && errno == EBADF) + fd_kill (EV_A_ fd); +} + +/* called on ENOMEM in select/poll to kill some fds and retry */ +static void noinline +fd_enomem (EV_P) +{ + int fd; + + for (fd = anfdmax; fd--; ) + if (anfds [fd].events) + { + fd_kill (EV_A_ fd); + return; + } +} + +/* usually called after fork if backend needs to re-arm all fds from scratch */ +static void noinline +fd_rearm_all (EV_P) +{ + int fd; + + for (fd = 0; fd < anfdmax; ++fd) + if (anfds [fd].events) + { + anfds [fd].events = 0; + fd_change (EV_A_ fd, EV_IOFDSET | 1); + } +} + +/*****************************************************************************/ + +void inline_speed +upheap (WT *heap, int k) +{ + WT w = heap [k]; + + while (k) + { + int p = (k - 1) >> 1; + + if (heap [p]->at <= w->at) + break; + + heap [k] = heap [p]; + ((W)heap [k])->active = k + 1; + k = p; + } + + heap [k] = w; + ((W)heap [k])->active = k + 1; +} + +void inline_speed +downheap (WT *heap, int N, int k) { - WT w = timers [k]; + WT w = heap [k]; - while (k < (N >> 1)) + for (;;) { - int j = k << 1; + int c = (k << 1) + 1; + + if (c >= N) + break; - if (j + 1 < N && timers [j]->at > timers [j + 1]->at) - ++j; + c += c + 1 < N && heap [c]->at > heap [c + 1]->at + ? 1 : 0; - if (w->at <= timers [j]->at) + if (w->at <= heap [c]->at) break; - timers [k] = timers [j]; - timers [k]->active = k + 1; - k = j; + heap [k] = heap [c]; + ((W)heap [k])->active = k + 1; + + k = c; } - timers [k] = w; - timers [k]->active = k + 1; + heap [k] = w; + ((W)heap [k])->active = k + 1; +} + +void inline_size +adjustheap (WT *heap, int N, int k) +{ + upheap (heap, k); + downheap (heap, N, k); } /*****************************************************************************/ typedef struct { - struct ev_signal *head; - sig_atomic_t gotsig; + WL head; + sig_atomic_t volatile gotsig; } ANSIG; static ANSIG *signals; static int signalmax; static int sigpipe [2]; -static sig_atomic_t gotsig; -static struct ev_io sigev; +static sig_atomic_t volatile gotsig; +static ev_io sigev; -static void +void inline_size signals_init (ANSIG *base, int count) { while (count--) { base->head = 0; base->gotsig = 0; + ++base; } } @@ -248,620 +790,1666 @@ signals_init (ANSIG *base, int count) static void sighandler (int signum) { - signals [signum - 1].gotsig = 1; +#if _WIN32 + signal (signum, sighandler); +#endif + + signals [signum - 1].gotsig = 1; if (!gotsig) { + int old_errno = errno; gotsig = 1; - write (sigpipe [1], &gotsig, 1); + write (sigpipe [1], &signum, 1); + errno = old_errno; } } +void noinline +ev_feed_signal_event (EV_P_ int signum) +{ + WL w; + +#if EV_MULTIPLICITY + assert (("feeding signal events is only supported in the default loop", loop == ev_default_loop_ptr)); +#endif + + --signum; + + if (signum < 0 || signum >= signalmax) + return; + + signals [signum].gotsig = 0; + + for (w = signals [signum].head; w; w = w->next) + ev_feed_event (EV_A_ (W)w, EV_SIGNAL); +} + static void -sigcb (struct ev_io *iow, int revents) +sigcb (EV_P_ ev_io *iow, int revents) { - struct ev_signal *w; - int sig; + int signum; - gotsig = 0; read (sigpipe [0], &revents, 1); + gotsig = 0; - for (sig = signalmax; sig--; ) - if (signals [sig].gotsig) - { - signals [sig].gotsig = 0; + for (signum = signalmax; signum--; ) + if (signals [signum].gotsig) + ev_feed_signal_event (EV_A_ signum + 1); +} + +void inline_speed +fd_intern (int fd) +{ +#ifdef _WIN32 + int arg = 1; + ioctlsocket (_get_osfhandle (fd), FIONBIO, &arg); +#else + fcntl (fd, F_SETFD, FD_CLOEXEC); + fcntl (fd, F_SETFL, O_NONBLOCK); +#endif +} + +static void noinline +siginit (EV_P) +{ + fd_intern (sigpipe [0]); + fd_intern (sigpipe [1]); + + ev_io_set (&sigev, sigpipe [0], EV_READ); + ev_io_start (EV_A_ &sigev); + ev_unref (EV_A); /* child watcher should not keep loop alive */ +} + +/*****************************************************************************/ + +static WL childs [EV_PID_HASHSIZE]; - for (w = signals [sig].head; w; w = w->next) - event ((W)w, EV_SIGNAL); +#ifndef _WIN32 + +static ev_signal childev; + +void inline_speed +child_reap (EV_P_ ev_signal *sw, int chain, int pid, int status) +{ + ev_child *w; + + for (w = (ev_child *)childs [chain & (EV_PID_HASHSIZE - 1)]; w; w = (ev_child *)((WL)w)->next) + if (w->pid == pid || !w->pid) + { + ev_set_priority (w, ev_priority (sw)); /* need to do it *now* */ + w->rpid = pid; + w->rstatus = status; + ev_feed_event (EV_A_ (W)w, EV_CHILD); } } +#ifndef WCONTINUED +# define WCONTINUED 0 +#endif + static void -siginit (void) +childcb (EV_P_ ev_signal *sw, int revents) +{ + int pid, status; + + /* some systems define WCONTINUED but then fail to support it (linux 2.4) */ + if (0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED))) + if (!WCONTINUED + || errno != EINVAL + || 0 >= (pid = waitpid (-1, &status, WNOHANG | WUNTRACED))) + return; + + /* make sure we are called again until all childs have been reaped */ + /* we need to do it this way so that the callback gets called before we continue */ + ev_feed_event (EV_A_ (W)sw, EV_SIGNAL); + + child_reap (EV_A_ sw, pid, pid, status); + if (EV_PID_HASHSIZE > 1) + child_reap (EV_A_ sw, 0, pid, status); /* this might trigger a watcher twice, but feed_event catches that */ +} + +#endif + +/*****************************************************************************/ + +#if EV_USE_PORT +# include "ev_port.c" +#endif +#if EV_USE_KQUEUE +# include "ev_kqueue.c" +#endif +#if EV_USE_EPOLL +# include "ev_epoll.c" +#endif +#if EV_USE_POLL +# include "ev_poll.c" +#endif +#if EV_USE_SELECT +# include "ev_select.c" +#endif + +int +ev_version_major (void) +{ + return EV_VERSION_MAJOR; +} + +int +ev_version_minor (void) +{ + return EV_VERSION_MINOR; +} + +/* return true if we are running with elevated privileges and should ignore env variables */ +int inline_size +enable_secure (void) +{ +#ifdef _WIN32 + return 0; +#else + return getuid () != geteuid () + || getgid () != getegid (); +#endif +} + +unsigned int +ev_supported_backends (void) +{ + unsigned int flags = 0; + + if (EV_USE_PORT ) flags |= EVBACKEND_PORT; + if (EV_USE_KQUEUE) flags |= EVBACKEND_KQUEUE; + if (EV_USE_EPOLL ) flags |= EVBACKEND_EPOLL; + if (EV_USE_POLL ) flags |= EVBACKEND_POLL; + if (EV_USE_SELECT) flags |= EVBACKEND_SELECT; + + return flags; +} + +unsigned int +ev_recommended_backends (void) +{ + unsigned int flags = ev_supported_backends (); + +#ifndef __NetBSD__ + /* kqueue is borked on everything but netbsd apparently */ + /* it usually doesn't work correctly on anything but sockets and pipes */ + flags &= ~EVBACKEND_KQUEUE; +#endif +#ifdef __APPLE__ + // flags &= ~EVBACKEND_KQUEUE; for documentation + flags &= ~EVBACKEND_POLL; +#endif + + return flags; +} + +unsigned int +ev_embeddable_backends (void) +{ + int flags = EVBACKEND_EPOLL | EVBACKEND_KQUEUE | EVBACKEND_PORT; + + /* epoll embeddability broken on all linux versions up to at least 2.6.23 */ + /* please fix it and tell me how to detect the fix */ + flags &= ~EVBACKEND_EPOLL; + + return flags; +} + +unsigned int +ev_backend (EV_P) +{ + return backend; +} + +unsigned int +ev_loop_count (EV_P) +{ + return loop_count; +} + +void +ev_set_io_collect_interval (EV_P_ ev_tstamp interval) +{ + io_blocktime = interval; +} + +void +ev_set_timeout_collect_interval (EV_P_ ev_tstamp interval) +{ + timeout_blocktime = interval; +} + +static void noinline +loop_init (EV_P_ unsigned int flags) +{ + if (!backend) + { +#if EV_USE_MONOTONIC + { + struct timespec ts; + if (!clock_gettime (CLOCK_MONOTONIC, &ts)) + have_monotonic = 1; + } +#endif + + ev_rt_now = ev_time (); + mn_now = get_clock (); + now_floor = mn_now; + rtmn_diff = ev_rt_now - mn_now; + + io_blocktime = 0.; + timeout_blocktime = 0.; + + /* pid check not overridable via env */ +#ifndef _WIN32 + if (flags & EVFLAG_FORKCHECK) + curpid = getpid (); +#endif + + if (!(flags & EVFLAG_NOENV) + && !enable_secure () + && getenv ("LIBEV_FLAGS")) + flags = atoi (getenv ("LIBEV_FLAGS")); + + if (!(flags & 0x0000ffffUL)) + flags |= ev_recommended_backends (); + + backend = 0; + backend_fd = -1; +#if EV_USE_INOTIFY + fs_fd = -2; +#endif + +#if EV_USE_PORT + if (!backend && (flags & EVBACKEND_PORT )) backend = port_init (EV_A_ flags); +#endif +#if EV_USE_KQUEUE + if (!backend && (flags & EVBACKEND_KQUEUE)) backend = kqueue_init (EV_A_ flags); +#endif +#if EV_USE_EPOLL + if (!backend && (flags & EVBACKEND_EPOLL )) backend = epoll_init (EV_A_ flags); +#endif +#if EV_USE_POLL + if (!backend && (flags & EVBACKEND_POLL )) backend = poll_init (EV_A_ flags); +#endif +#if EV_USE_SELECT + if (!backend && (flags & EVBACKEND_SELECT)) backend = select_init (EV_A_ flags); +#endif + + ev_init (&sigev, sigcb); + ev_set_priority (&sigev, EV_MAXPRI); + } +} + +static void noinline +loop_destroy (EV_P) +{ + int i; + +#if EV_USE_INOTIFY + if (fs_fd >= 0) + close (fs_fd); +#endif + + if (backend_fd >= 0) + close (backend_fd); + +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_destroy (EV_A); +#endif +#if EV_USE_KQUEUE + if (backend == EVBACKEND_KQUEUE) kqueue_destroy (EV_A); +#endif +#if EV_USE_EPOLL + if (backend == EVBACKEND_EPOLL ) epoll_destroy (EV_A); +#endif +#if EV_USE_POLL + if (backend == EVBACKEND_POLL ) poll_destroy (EV_A); +#endif +#if EV_USE_SELECT + if (backend == EVBACKEND_SELECT) select_destroy (EV_A); +#endif + + for (i = NUMPRI; i--; ) + { + array_free (pending, [i]); +#if EV_IDLE_ENABLE + array_free (idle, [i]); +#endif + } + + ev_free (anfds); anfdmax = 0; + + /* have to use the microsoft-never-gets-it-right macro */ + array_free (fdchange, EMPTY); + array_free (timer, EMPTY); +#if EV_PERIODIC_ENABLE + array_free (periodic, EMPTY); +#endif +#if EV_FORK_ENABLE + array_free (fork, EMPTY); +#endif + array_free (prepare, EMPTY); + array_free (check, EMPTY); + + backend = 0; +} + +void inline_size infy_fork (EV_P); + +void inline_size +loop_fork (EV_P) +{ +#if EV_USE_PORT + if (backend == EVBACKEND_PORT ) port_fork (EV_A); +#endif +#if EV_USE_KQUEUE + if (backend == EVBACKEND_KQUEUE) kqueue_fork (EV_A); +#endif +#if EV_USE_EPOLL + if (backend == EVBACKEND_EPOLL ) epoll_fork (EV_A); +#endif +#if EV_USE_INOTIFY + infy_fork (EV_A); +#endif + + if (ev_is_active (&sigev)) + { + /* default loop */ + + ev_ref (EV_A); + ev_io_stop (EV_A_ &sigev); + close (sigpipe [0]); + close (sigpipe [1]); + + while (pipe (sigpipe)) + syserr ("(libev) error creating pipe"); + + siginit (EV_A); + sigcb (EV_A_ &sigev, EV_READ); + } + + postfork = 0; +} + +#if EV_MULTIPLICITY +struct ev_loop * +ev_loop_new (unsigned int flags) +{ + struct ev_loop *loop = (struct ev_loop *)ev_malloc (sizeof (struct ev_loop)); + + memset (loop, 0, sizeof (struct ev_loop)); + + loop_init (EV_A_ flags); + + if (ev_backend (EV_A)) + return loop; + + return 0; +} + +void +ev_loop_destroy (EV_P) +{ + loop_destroy (EV_A); + ev_free (loop); +} + +void +ev_loop_fork (EV_P) +{ + postfork = 1; // must be in line with ev_default_fork +} + +#endif + +#if EV_MULTIPLICITY +struct ev_loop * +ev_default_loop_init (unsigned int flags) +#else +int +ev_default_loop (unsigned int flags) +#endif +{ + if (sigpipe [0] == sigpipe [1]) + if (pipe (sigpipe)) + return 0; + + if (!ev_default_loop_ptr) + { +#if EV_MULTIPLICITY + struct ev_loop *loop = ev_default_loop_ptr = &default_loop_struct; +#else + ev_default_loop_ptr = 1; +#endif + + loop_init (EV_A_ flags); + + if (ev_backend (EV_A)) + { + siginit (EV_A); + +#ifndef _WIN32 + ev_signal_init (&childev, childcb, SIGCHLD); + ev_set_priority (&childev, EV_MAXPRI); + ev_signal_start (EV_A_ &childev); + ev_unref (EV_A); /* child watcher should not keep loop alive */ +#endif + } + else + ev_default_loop_ptr = 0; + } + + return ev_default_loop_ptr; +} + +void +ev_default_destroy (void) { - fcntl (sigpipe [0], F_SETFD, FD_CLOEXEC); - fcntl (sigpipe [1], F_SETFD, FD_CLOEXEC); +#if EV_MULTIPLICITY + struct ev_loop *loop = ev_default_loop_ptr; +#endif + +#ifndef _WIN32 + ev_ref (EV_A); /* child watcher */ + ev_signal_stop (EV_A_ &childev); +#endif + + ev_ref (EV_A); /* signal watcher */ + ev_io_stop (EV_A_ &sigev); + + close (sigpipe [0]); sigpipe [0] = 0; + close (sigpipe [1]); sigpipe [1] = 0; + + loop_destroy (EV_A); +} - /* rather than sort out wether we really need nb, set it */ - fcntl (sigpipe [0], F_SETFL, O_NONBLOCK); - fcntl (sigpipe [1], F_SETFL, O_NONBLOCK); +void +ev_default_fork (void) +{ +#if EV_MULTIPLICITY + struct ev_loop *loop = ev_default_loop_ptr; +#endif - evio_set (&sigev, sigpipe [0], EV_READ); - evio_start (&sigev); + if (backend) + postfork = 1; // must be in line with ev_loop_fork } /*****************************************************************************/ -static struct ev_idle **idles; -static int idlemax, idlecnt; +void +ev_invoke (EV_P_ void *w, int revents) +{ + EV_CB_INVOKE ((W)w, revents); +} + +void inline_speed +call_pending (EV_P) +{ + int pri; + + for (pri = NUMPRI; pri--; ) + while (pendingcnt [pri]) + { + ANPENDING *p = pendings [pri] + --pendingcnt [pri]; + + if (expect_true (p->w)) + { + /*assert (("non-pending watcher on pending list", p->w->pending));*/ + + p->w->pending = 0; + EV_CB_INVOKE (p->w, p->events); + } + } +} + +void inline_size +timers_reify (EV_P) +{ + while (timercnt && ((WT)timers [0])->at <= mn_now) + { + ev_timer *w = (ev_timer *)timers [0]; + + /*assert (("inactive timer on timer heap detected", ev_is_active (w)));*/ + + /* first reschedule or stop timer */ + if (w->repeat) + { + assert (("negative ev_timer repeat value found while processing timers", w->repeat > 0.)); -static struct ev_check **checks; -static int checkmax, checkcnt; + ((WT)w)->at += w->repeat; + if (((WT)w)->at < mn_now) + ((WT)w)->at = mn_now; + + downheap (timers, timercnt, 0); + } + else + ev_timer_stop (EV_A_ w); /* nonrepeating: stop timer */ + + ev_feed_event (EV_A_ (W)w, EV_TIMEOUT); + } +} + +#if EV_PERIODIC_ENABLE +void inline_size +periodics_reify (EV_P) +{ + while (periodiccnt && ((WT)periodics [0])->at <= ev_rt_now) + { + ev_periodic *w = (ev_periodic *)periodics [0]; + + /*assert (("inactive timer on periodic heap detected", ev_is_active (w)));*/ + + /* first reschedule or stop timer */ + if (w->reschedule_cb) + { + ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + TIME_EPSILON); + assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > ev_rt_now)); + downheap (periodics, periodiccnt, 0); + } + else if (w->interval) + { + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + if (((WT)w)->at - ev_rt_now <= TIME_EPSILON) ((WT)w)->at += w->interval; + assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > ev_rt_now)); + downheap (periodics, periodiccnt, 0); + } + else + ev_periodic_stop (EV_A_ w); /* nonrepeating: stop timer */ + + ev_feed_event (EV_A_ (W)w, EV_PERIODIC); + } +} + +static void noinline +periodics_reschedule (EV_P) +{ + int i; + + /* adjust periodics after time jump */ + for (i = 0; i < periodiccnt; ++i) + { + ev_periodic *w = (ev_periodic *)periodics [i]; + + if (w->reschedule_cb) + ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); + else if (w->interval) + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + } + + /* now rebuild the heap */ + for (i = periodiccnt >> 1; i--; ) + downheap (periodics, periodiccnt, i); +} +#endif + +#if EV_IDLE_ENABLE +void inline_size +idle_reify (EV_P) +{ + if (expect_false (idleall)) + { + int pri; + + for (pri = NUMPRI; pri--; ) + { + if (pendingcnt [pri]) + break; + + if (idlecnt [pri]) + { + queue_events (EV_A_ (W *)idles [pri], idlecnt [pri], EV_IDLE); + break; + } + } + } +} +#endif + +void inline_speed +time_update (EV_P_ ev_tstamp max_block) +{ + int i; + +#if EV_USE_MONOTONIC + if (expect_true (have_monotonic)) + { + ev_tstamp odiff = rtmn_diff; + + mn_now = get_clock (); + + /* only fetch the realtime clock every 0.5*MIN_TIMEJUMP seconds */ + /* interpolate in the meantime */ + if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5)) + { + ev_rt_now = rtmn_diff + mn_now; + return; + } + + now_floor = mn_now; + ev_rt_now = ev_time (); + + /* loop a few times, before making important decisions. + * on the choice of "4": one iteration isn't enough, + * in case we get preempted during the calls to + * ev_time and get_clock. a second call is almost guaranteed + * to succeed in that case, though. and looping a few more times + * doesn't hurt either as we only do this on time-jumps or + * in the unlikely event of having been preempted here. + */ + for (i = 4; --i; ) + { + rtmn_diff = ev_rt_now - mn_now; + + if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP) + return; /* all is well */ + + ev_rt_now = ev_time (); + mn_now = get_clock (); + now_floor = mn_now; + } + +# if EV_PERIODIC_ENABLE + periodics_reschedule (EV_A); +# endif + /* no timer adjustment, as the monotonic clock doesn't jump */ + /* timers_reschedule (EV_A_ rtmn_diff - odiff) */ + } + else +#endif + { + ev_rt_now = ev_time (); + + if (expect_false (mn_now > ev_rt_now || ev_rt_now > mn_now + max_block + MIN_TIMEJUMP)) + { +#if EV_PERIODIC_ENABLE + periodics_reschedule (EV_A); +#endif + /* adjust timers. this is easy, as the offset is the same for all of them */ + for (i = 0; i < timercnt; ++i) + ((WT)timers [i])->at += ev_rt_now - mn_now; + } + + mn_now = ev_rt_now; + } +} + +void +ev_ref (EV_P) +{ + ++activecnt; +} + +void +ev_unref (EV_P) +{ + --activecnt; +} + +static int loop_done; + +void +ev_loop (EV_P_ int flags) +{ + loop_done = flags & (EVLOOP_ONESHOT | EVLOOP_NONBLOCK) + ? EVUNLOOP_ONE + : EVUNLOOP_CANCEL; + + call_pending (EV_A); /* in case we recurse, ensure ordering stays nice and clean */ + + do + { +#ifndef _WIN32 + if (expect_false (curpid)) /* penalise the forking check even more */ + if (expect_false (getpid () != curpid)) + { + curpid = getpid (); + postfork = 1; + } +#endif + +#if EV_FORK_ENABLE + /* we might have forked, so queue fork handlers */ + if (expect_false (postfork)) + if (forkcnt) + { + queue_events (EV_A_ (W *)forks, forkcnt, EV_FORK); + call_pending (EV_A); + } +#endif + + /* queue prepare watchers (and execute them) */ + if (expect_false (preparecnt)) + { + queue_events (EV_A_ (W *)prepares, preparecnt, EV_PREPARE); + call_pending (EV_A); + } + + if (expect_false (!activecnt)) + break; + + /* we might have forked, so reify kernel state if necessary */ + if (expect_false (postfork)) + loop_fork (EV_A); + + /* update fd-related kernel structures */ + fd_reify (EV_A); + + /* calculate blocking time */ + { + ev_tstamp waittime = 0.; + ev_tstamp sleeptime = 0.; + + if (expect_true (!(flags & EVLOOP_NONBLOCK || idleall || !activecnt))) + { + /* update time to cancel out callback processing overhead */ + time_update (EV_A_ 1e100); + + waittime = MAX_BLOCKTIME; + + if (timercnt) + { + ev_tstamp to = ((WT)timers [0])->at - mn_now + backend_fudge; + if (waittime > to) waittime = to; + } + +#if EV_PERIODIC_ENABLE + if (periodiccnt) + { + ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + backend_fudge; + if (waittime > to) waittime = to; + } +#endif + + if (expect_false (waittime < timeout_blocktime)) + waittime = timeout_blocktime; + + sleeptime = waittime - backend_fudge; + + if (expect_true (sleeptime > io_blocktime)) + sleeptime = io_blocktime; + + if (sleeptime) + { + ev_sleep (sleeptime); + waittime -= sleeptime; + } + } + + ++loop_count; + backend_poll (EV_A_ waittime); + + /* update ev_rt_now, do magic */ + time_update (EV_A_ waittime + sleeptime); + } + + /* queue pending timers and reschedule them */ + timers_reify (EV_A); /* relative timers called last */ +#if EV_PERIODIC_ENABLE + periodics_reify (EV_A); /* absolute timers called first */ +#endif + +#if EV_IDLE_ENABLE + /* queue idle watchers unless other events are pending */ + idle_reify (EV_A); +#endif + + /* queue check watchers, to be executed first */ + if (expect_false (checkcnt)) + queue_events (EV_A_ (W *)checks, checkcnt, EV_CHECK); + + call_pending (EV_A); + + } + while (expect_true (activecnt && !loop_done)); + + if (loop_done == EVUNLOOP_ONE) + loop_done = EVUNLOOP_CANCEL; +} + +void +ev_unloop (EV_P_ int how) +{ + loop_done = how; +} /*****************************************************************************/ -#if HAVE_EPOLL -# include "ev_epoll.c" +void inline_size +wlist_add (WL *head, WL elem) +{ + elem->next = *head; + *head = elem; +} + +void inline_size +wlist_del (WL *head, WL elem) +{ + while (*head) + { + if (*head == elem) + { + *head = elem->next; + return; + } + + head = &(*head)->next; + } +} + +void inline_speed +clear_pending (EV_P_ W w) +{ + if (w->pending) + { + pendings [ABSPRI (w)][w->pending - 1].w = 0; + w->pending = 0; + } +} + +int +ev_clear_pending (EV_P_ void *w) +{ + W w_ = (W)w; + int pending = w_->pending; + + if (expect_true (pending)) + { + ANPENDING *p = pendings [ABSPRI (w_)] + pending - 1; + w_->pending = 0; + p->w = 0; + return p->events; + } + else + return 0; +} + +void inline_size +pri_adjust (EV_P_ W w) +{ + int pri = w->priority; + pri = pri < EV_MINPRI ? EV_MINPRI : pri; + pri = pri > EV_MAXPRI ? EV_MAXPRI : pri; + w->priority = pri; +} + +void inline_speed +ev_start (EV_P_ W w, int active) +{ + pri_adjust (EV_A_ w); + w->active = active; + ev_ref (EV_A); +} + +void inline_size +ev_stop (EV_P_ W w) +{ + ev_unref (EV_A); + w->active = 0; +} + +/*****************************************************************************/ + +void noinline +ev_io_start (EV_P_ ev_io *w) +{ + int fd = w->fd; + + if (expect_false (ev_is_active (w))) + return; + + assert (("ev_io_start called with negative fd", fd >= 0)); + + ev_start (EV_A_ (W)w, 1); + array_needsize (ANFD, anfds, anfdmax, fd + 1, anfds_init); + wlist_add (&anfds[fd].head, (WL)w); + + fd_change (EV_A_ fd, w->events & EV_IOFDSET | 1); + w->events &= ~EV_IOFDSET; +} + +void noinline +ev_io_stop (EV_P_ ev_io *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + assert (("ev_io_start called with illegal fd (must stay constant after start!)", w->fd >= 0 && w->fd < anfdmax)); + + wlist_del (&anfds[w->fd].head, (WL)w); + ev_stop (EV_A_ (W)w); + + fd_change (EV_A_ w->fd, 1); +} + +void noinline +ev_timer_start (EV_P_ ev_timer *w) +{ + if (expect_false (ev_is_active (w))) + return; + + ((WT)w)->at += mn_now; + + assert (("ev_timer_start called with negative timer repeat value", w->repeat >= 0.)); + + ev_start (EV_A_ (W)w, ++timercnt); + array_needsize (WT, timers, timermax, timercnt, EMPTY2); + timers [timercnt - 1] = (WT)w; + upheap (timers, timercnt - 1); + + /*assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w));*/ +} + +void noinline +ev_timer_stop (EV_P_ ev_timer *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + assert (("internal timer heap corruption", timers [((W)w)->active - 1] == (WT)w)); + + { + int active = ((W)w)->active; + + if (expect_true (--active < --timercnt)) + { + timers [active] = timers [timercnt]; + adjustheap (timers, timercnt, active); + } + } + + ((WT)w)->at -= mn_now; + + ev_stop (EV_A_ (W)w); +} + +void noinline +ev_timer_again (EV_P_ ev_timer *w) +{ + if (ev_is_active (w)) + { + if (w->repeat) + { + ((WT)w)->at = mn_now + w->repeat; + adjustheap (timers, timercnt, ((W)w)->active - 1); + } + else + ev_timer_stop (EV_A_ w); + } + else if (w->repeat) + { + w->at = w->repeat; + ev_timer_start (EV_A_ w); + } +} + +#if EV_PERIODIC_ENABLE +void noinline +ev_periodic_start (EV_P_ ev_periodic *w) +{ + if (expect_false (ev_is_active (w))) + return; + + if (w->reschedule_cb) + ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); + else if (w->interval) + { + assert (("ev_periodic_start called with negative interval value", w->interval >= 0.)); + /* this formula differs from the one in periodic_reify because we do not always round up */ + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + } + else + ((WT)w)->at = w->offset; + + ev_start (EV_A_ (W)w, ++periodiccnt); + array_needsize (WT, periodics, periodicmax, periodiccnt, EMPTY2); + periodics [periodiccnt - 1] = (WT)w; + upheap (periodics, periodiccnt - 1); + + /*assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w));*/ +} + +void noinline +ev_periodic_stop (EV_P_ ev_periodic *w) +{ + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == (WT)w)); + + { + int active = ((W)w)->active; + + if (expect_true (--active < --periodiccnt)) + { + periodics [active] = periodics [periodiccnt]; + adjustheap (periodics, periodiccnt, active); + } + } + + ev_stop (EV_A_ (W)w); +} + +void noinline +ev_periodic_again (EV_P_ ev_periodic *w) +{ + /* TODO: use adjustheap and recalculation */ + ev_periodic_stop (EV_A_ w); + ev_periodic_start (EV_A_ w); +} #endif -#if HAVE_SELECT -# include "ev_select.c" + +#ifndef SA_RESTART +# define SA_RESTART 0 #endif -int ev_init (int flags) +void noinline +ev_signal_start (EV_P_ ev_signal *w) { -#if HAVE_MONOTONIC - { - struct timespec ts; - if (!clock_gettime (CLOCK_MONOTONIC, &ts)) - have_monotonic = 1; - } +#if EV_MULTIPLICITY + assert (("signal watchers are only supported in the default loop", loop == ev_default_loop_ptr)); #endif + if (expect_false (ev_is_active (w))) + return; - ev_now = ev_time (); - now = get_clock (); - diff = ev_now - now; - - if (pipe (sigpipe)) - return 0; + assert (("ev_signal_start called with illegal signal number", w->signum > 0)); - ev_method = EVMETHOD_NONE; -#if HAVE_EPOLL - if (ev_method == EVMETHOD_NONE) epoll_init (flags); + { +#ifndef _WIN32 + sigset_t full, prev; + sigfillset (&full); + sigprocmask (SIG_SETMASK, &full, &prev); #endif -#if HAVE_SELECT - if (ev_method == EVMETHOD_NONE) select_init (flags); + + array_needsize (ANSIG, signals, signalmax, w->signum, signals_init); + +#ifndef _WIN32 + sigprocmask (SIG_SETMASK, &prev, 0); #endif + } - if (ev_method) + ev_start (EV_A_ (W)w, 1); + wlist_add (&signals [w->signum - 1].head, (WL)w); + + if (!((WL)w)->next) { - evw_init (&sigev, sigcb); - siginit (); +#if _WIN32 + signal (w->signum, sighandler); +#else + struct sigaction sa; + sa.sa_handler = sighandler; + sigfillset (&sa.sa_mask); + sa.sa_flags = SA_RESTART; /* if restarting works we save one iteration */ + sigaction (w->signum, &sa, 0); +#endif } - - return ev_method; } -/*****************************************************************************/ - -void ev_prefork (void) +void noinline +ev_signal_stop (EV_P_ ev_signal *w) { - /* nop */ -} + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; -void ev_postfork_parent (void) -{ - /* nop */ + wlist_del (&signals [w->signum - 1].head, (WL)w); + ev_stop (EV_A_ (W)w); + + if (!signals [w->signum - 1].head) + signal (w->signum, SIG_DFL); } -void ev_postfork_child (void) +void +ev_child_start (EV_P_ ev_child *w) { -#if HAVE_EPOLL - if (ev_method == EVMETHOD_EPOLL) - epoll_postfork_child (); +#if EV_MULTIPLICITY + assert (("child watchers are only supported in the default loop", loop == ev_default_loop_ptr)); #endif + if (expect_false (ev_is_active (w))) + return; - evio_stop (&sigev); - close (sigpipe [0]); - close (sigpipe [1]); - pipe (sigpipe); - siginit (); + ev_start (EV_A_ (W)w, 1); + wlist_add (&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); } -/*****************************************************************************/ - -static void -fd_reify (void) +void +ev_child_stop (EV_P_ ev_child *w) { - int i; + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; - for (i = 0; i < fdchangecnt; ++i) - { - int fd = fdchanges [i]; - ANFD *anfd = anfds + fd; - struct ev_io *w; + wlist_del (&childs [w->pid & (EV_PID_HASHSIZE - 1)], (WL)w); + ev_stop (EV_A_ (W)w); +} - int wev = 0; +#if EV_STAT_ENABLE - for (w = anfd->head; w; w = w->next) - wev |= w->events; +# ifdef _WIN32 +# undef lstat +# define lstat(a,b) _stati64 (a,b) +# endif - if (anfd->wev != wev) - { - method_modify (fd, anfd->wev, wev); - anfd->wev = wev; - } - } +#define DEF_STAT_INTERVAL 5.0074891 +#define MIN_STAT_INTERVAL 0.1074891 - fdchangecnt = 0; -} +static void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents); -static void -call_pending () +#if EV_USE_INOTIFY +# define EV_INOTIFY_BUFSIZE 8192 + +static void noinline +infy_add (EV_P_ ev_stat *w) { - int i; + w->wd = inotify_add_watch (fs_fd, w->path, IN_ATTRIB | IN_DELETE_SELF | IN_MOVE_SELF | IN_MODIFY | IN_DONT_FOLLOW | IN_MASK_ADD); - for (i = 0; i < pendingcnt; ++i) + if (w->wd < 0) { - ANPENDING *p = pendings + i; + ev_timer_start (EV_A_ &w->timer); /* this is not race-free, so we still need to recheck periodically */ - if (p->w) + /* monitor some parent directory for speedup hints */ + if ((errno == ENOENT || errno == EACCES) && strlen (w->path) < 4096) { - p->w->pending = 0; - p->w->cb (p->w, p->events); - } - } + char path [4096]; + strcpy (path, w->path); - pendingcnt = 0; -} + do + { + int mask = IN_MASK_ADD | IN_DELETE_SELF | IN_MOVE_SELF + | (errno == EACCES ? IN_ATTRIB : IN_CREATE | IN_MOVED_TO); -static void -timers_reify () -{ - while (timercnt && timers [0]->at <= now) - { - struct ev_timer *w = timers [0]; + char *pend = strrchr (path, '/'); - /* first reschedule or stop timer */ - if (w->repeat) - { - w->at = now + w->repeat; - assert (("timer timeout in the past, negative repeat?", w->at > now)); - downheap ((WT *)timers, timercnt, 0); - } - else - evtimer_stop (w); /* nonrepeating: stop timer */ + if (!pend) + break; /* whoops, no '/', complain to your admin */ - event ((W)w, EV_TIMEOUT); + *pend = 0; + w->wd = inotify_add_watch (fs_fd, path, mask); + } + while (w->wd < 0 && (errno == ENOENT || errno == EACCES)); + } } + else + ev_timer_stop (EV_A_ &w->timer); /* we can watch this in a race-free way */ + + if (w->wd >= 0) + wlist_add (&fs_hash [w->wd & (EV_INOTIFY_HASHSIZE - 1)].head, (WL)w); } -static void -periodics_reify () +static void noinline +infy_del (EV_P_ ev_stat *w) { - while (periodiccnt && periodics [0]->at <= ev_now) - { - struct ev_periodic *w = periodics [0]; + int slot; + int wd = w->wd; - /* first reschedule or stop timer */ - if (w->interval) - { - w->at += floor ((ev_now - w->at) / w->interval + 1.) * w->interval; - assert (("periodic timeout in the past, negative interval?", w->at > ev_now)); - downheap ((WT *)periodics, periodiccnt, 0); - } - else - evperiodic_stop (w); /* nonrepeating: stop timer */ + if (wd < 0) + return; - event ((W)w, EV_TIMEOUT); - } + w->wd = -2; + slot = wd & (EV_INOTIFY_HASHSIZE - 1); + wlist_del (&fs_hash [slot].head, (WL)w); + + /* remove this watcher, if others are watching it, they will rearm */ + inotify_rm_watch (fs_fd, wd); } -static void -time_jump (ev_tstamp diff) +static void noinline +infy_wd (EV_P_ int slot, int wd, struct inotify_event *ev) { - int i; - - /* adjust periodics */ - for (i = 0; i < periodiccnt; ++i) + if (slot < 0) + /* overflow, need to check for all hahs slots */ + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + infy_wd (EV_A_ slot, wd, ev); + else { - struct ev_periodic *w = periodics [i]; + WL w_; - if (w->interval) + for (w_ = fs_hash [slot & (EV_INOTIFY_HASHSIZE - 1)].head; w_; ) { - ev_tstamp diff = ceil ((ev_now - w->at) / w->interval) * w->interval; + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us remove this watcher and all before it */ - if (fabs (diff) >= 1e-4) + if (w->wd == wd || wd == -1) { - evperiodic_stop (w); - evperiodic_start (w); + if (ev->mask & (IN_IGNORED | IN_UNMOUNT | IN_DELETE_SELF)) + { + w->wd = -1; + infy_add (EV_A_ w); /* re-add, no matter what */ + } - i = 0; /* restart loop, inefficient, but time jumps should be rare */ + stat_timer_cb (EV_A_ &w->timer, 0); } } } - - /* adjust timers. this is easy, as the offset is the same for all */ - for (i = 0; i < timercnt; ++i) - timers [i]->at += diff; } static void -time_update () +infy_cb (EV_P_ ev_io *w, int revents) { - int i; - - ev_now = ev_time (); - - if (have_monotonic) - { - ev_tstamp odiff = diff; + char buf [EV_INOTIFY_BUFSIZE]; + struct inotify_event *ev = (struct inotify_event *)buf; + int ofs; + int len = read (fs_fd, buf, sizeof (buf)); - for (i = 4; --i; ) /* loop a few times, before making important decisions */ - { - now = get_clock (); - diff = ev_now - now; + for (ofs = 0; ofs < len; ofs += sizeof (struct inotify_event) + ev->len) + infy_wd (EV_A_ ev->wd, ev->wd, ev); +} - if (fabs (odiff - diff) < MIN_TIMEJUMP) - return; /* all is well */ +void inline_size +infy_init (EV_P) +{ + if (fs_fd != -2) + return; - ev_now = ev_time (); - } + fs_fd = inotify_init (); - time_jump (diff - odiff); - } - else + if (fs_fd >= 0) { - if (now > ev_now || now < ev_now - MAX_BLOCKTIME - MIN_TIMEJUMP) - time_jump (ev_now - now); - - now = ev_now; + ev_io_init (&fs_w, infy_cb, fs_fd, EV_READ); + ev_set_priority (&fs_w, EV_MAXPRI); + ev_io_start (EV_A_ &fs_w); } } -int ev_loop_done; - -void ev_loop (int flags) +void inline_size +infy_fork (EV_P) { - double block; - ev_loop_done = flags & EVLOOP_ONESHOT; - - if (checkcnt) - { - queue_events ((W *)checks, checkcnt, EV_CHECK); - call_pending (); - } + int slot; - do - { - /* update fd-related kernel structures */ - fd_reify (); + if (fs_fd < 0) + return; - /* calculate blocking time */ + close (fs_fd); + fs_fd = inotify_init (); - /* we only need this for !monotonic clock, but as we always have timers, we just calculate it every time */ - ev_now = ev_time (); + for (slot = 0; slot < EV_INOTIFY_HASHSIZE; ++slot) + { + WL w_ = fs_hash [slot].head; + fs_hash [slot].head = 0; - if (flags & EVLOOP_NONBLOCK || idlecnt) - block = 0.; - else + while (w_) { - block = MAX_BLOCKTIME; + ev_stat *w = (ev_stat *)w_; + w_ = w_->next; /* lets us add this watcher */ - if (timercnt) - { - ev_tstamp to = timers [0]->at - get_clock () + method_fudge; - if (block > to) block = to; - } - - if (periodiccnt) - { - ev_tstamp to = periodics [0]->at - ev_now + method_fudge; - if (block > to) block = to; - } + w->wd = -1; - if (block < 0.) block = 0.; + if (fs_fd >= 0) + infy_add (EV_A_ w); /* re-add, no matter what */ + else + ev_timer_start (EV_A_ &w->timer); } - method_poll (block); - - /* update ev_now, do magic */ - time_update (); - - /* queue pending timers and reschedule them */ - periodics_reify (); /* absolute timers first */ - timers_reify (); /* relative timers second */ - - /* queue idle watchers unless io or timers are pending */ - if (!pendingcnt) - queue_events ((W *)idles, idlecnt, EV_IDLE); - - /* queue check and possibly idle watchers */ - queue_events ((W *)checks, checkcnt, EV_CHECK); - - call_pending (); } - while (!ev_loop_done); } -/*****************************************************************************/ +#endif -static void -wlist_add (WL *head, WL elem) +void +ev_stat_stat (EV_P_ ev_stat *w) { - elem->next = *head; - *head = elem; + if (lstat (w->path, &w->attr) < 0) + w->attr.st_nlink = 0; + else if (!w->attr.st_nlink) + w->attr.st_nlink = 1; } -static void -wlist_del (WL *head, WL elem) +static void noinline +stat_timer_cb (EV_P_ ev_timer *w_, int revents) { - while (*head) - { - if (*head == elem) - { - *head = elem->next; - return; - } - - head = &(*head)->next; + ev_stat *w = (ev_stat *)(((char *)w_) - offsetof (ev_stat, timer)); + + /* we copy this here each the time so that */ + /* prev has the old value when the callback gets invoked */ + w->prev = w->attr; + ev_stat_stat (EV_A_ w); + + /* memcmp doesn't work on netbsd, they.... do stuff to their struct stat */ + if ( + w->prev.st_dev != w->attr.st_dev + || w->prev.st_ino != w->attr.st_ino + || w->prev.st_mode != w->attr.st_mode + || w->prev.st_nlink != w->attr.st_nlink + || w->prev.st_uid != w->attr.st_uid + || w->prev.st_gid != w->attr.st_gid + || w->prev.st_rdev != w->attr.st_rdev + || w->prev.st_size != w->attr.st_size + || w->prev.st_atime != w->attr.st_atime + || w->prev.st_mtime != w->attr.st_mtime + || w->prev.st_ctime != w->attr.st_ctime + ) { + #if EV_USE_INOTIFY + infy_del (EV_A_ w); + infy_add (EV_A_ w); + ev_stat_stat (EV_A_ w); /* avoid race... */ + #endif + + ev_feed_event (EV_A_ w, EV_STAT); } } -static void -ev_start (W w, int active) +void +ev_stat_start (EV_P_ ev_stat *w) { - w->pending = 0; - w->active = active; + if (expect_false (ev_is_active (w))) + return; + + /* since we use memcmp, we need to clear any padding data etc. */ + memset (&w->prev, 0, sizeof (ev_statdata)); + memset (&w->attr, 0, sizeof (ev_statdata)); + + ev_stat_stat (EV_A_ w); + + if (w->interval < MIN_STAT_INTERVAL) + w->interval = w->interval ? MIN_STAT_INTERVAL : DEF_STAT_INTERVAL; + + ev_timer_init (&w->timer, stat_timer_cb, w->interval, w->interval); + ev_set_priority (&w->timer, ev_priority (w)); + +#if EV_USE_INOTIFY + infy_init (EV_A); + + if (fs_fd >= 0) + infy_add (EV_A_ w); + else +#endif + ev_timer_start (EV_A_ &w->timer); + + ev_start (EV_A_ (W)w, 1); } -static void -ev_stop (W w) +void +ev_stat_stop (EV_P_ ev_stat *w) { - if (w->pending) - pendings [w->pending - 1].w = 0; + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; - w->active = 0; -} +#if EV_USE_INOTIFY + infy_del (EV_A_ w); +#endif + ev_timer_stop (EV_A_ &w->timer); -/*****************************************************************************/ + ev_stop (EV_A_ (W)w); +} +#endif +#if EV_IDLE_ENABLE void -evio_start (struct ev_io *w) +ev_idle_start (EV_P_ ev_idle *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - int fd = w->fd; + pri_adjust (EV_A_ (W)w); + + { + int active = ++idlecnt [ABSPRI (w)]; - ev_start ((W)w, 1); - array_needsize (anfds, anfdmax, fd + 1, anfds_init); - wlist_add ((WL *)&anfds[fd].head, (WL)w); + ++idleall; + ev_start (EV_A_ (W)w, active); - ++fdchangecnt; - array_needsize (fdchanges, fdchangemax, fdchangecnt, ); - fdchanges [fdchangecnt - 1] = fd; + array_needsize (ev_idle *, idles [ABSPRI (w)], idlemax [ABSPRI (w)], active, EMPTY2); + idles [ABSPRI (w)][active - 1] = w; + } } void -evio_stop (struct ev_io *w) +ev_idle_stop (EV_P_ ev_idle *w) { - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - wlist_del ((WL *)&anfds[w->fd].head, (WL)w); - ev_stop ((W)w); + { + int active = ((W)w)->active; - ++fdchangecnt; - array_needsize (fdchanges, fdchangemax, fdchangecnt, ); - fdchanges [fdchangecnt - 1] = w->fd; -} + idles [ABSPRI (w)][active - 1] = idles [ABSPRI (w)][--idlecnt [ABSPRI (w)]]; + ((W)idles [ABSPRI (w)][active - 1])->active = active; + ev_stop (EV_A_ (W)w); + --idleall; + } +} +#endif void -evtimer_start (struct ev_timer *w) +ev_prepare_start (EV_P_ ev_prepare *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - w->at += now; - - ev_start ((W)w, ++timercnt); - array_needsize (timers, timermax, timercnt, ); - timers [timercnt - 1] = w; - upheap ((WT *)timers, timercnt - 1); + ev_start (EV_A_ (W)w, ++preparecnt); + array_needsize (ev_prepare *, prepares, preparemax, preparecnt, EMPTY2); + prepares [preparecnt - 1] = w; } void -evtimer_stop (struct ev_timer *w) +ev_prepare_stop (EV_P_ ev_prepare *w) { - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - if (w->active < timercnt--) - { - timers [w->active - 1] = timers [timercnt]; - downheap ((WT *)timers, timercnt, w->active - 1); - } + { + int active = ((W)w)->active; + prepares [active - 1] = prepares [--preparecnt]; + ((W)prepares [active - 1])->active = active; + } - ev_stop ((W)w); + ev_stop (EV_A_ (W)w); } void -evperiodic_start (struct ev_periodic *w) +ev_check_start (EV_P_ ev_check *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - /* this formula differs from the one in periodic_reify because we do not always round up */ - if (w->interval) - w->at += ceil ((ev_now - w->at) / w->interval) * w->interval; - - ev_start ((W)w, ++periodiccnt); - array_needsize (periodics, periodicmax, periodiccnt, ); - periodics [periodiccnt - 1] = w; - upheap ((WT *)periodics, periodiccnt - 1); + ev_start (EV_A_ (W)w, ++checkcnt); + array_needsize (ev_check *, checks, checkmax, checkcnt, EMPTY2); + checks [checkcnt - 1] = w; } void -evperiodic_stop (struct ev_periodic *w) +ev_check_stop (EV_P_ ev_check *w) { - if (!ev_is_active (w)) + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) return; - if (w->active < periodiccnt--) - { - periodics [w->active - 1] = periodics [periodiccnt]; - downheap ((WT *)periodics, periodiccnt, w->active - 1); - } + { + int active = ((W)w)->active; + checks [active - 1] = checks [--checkcnt]; + ((W)checks [active - 1])->active = active; + } - ev_stop ((W)w); + ev_stop (EV_A_ (W)w); } -void -evsignal_start (struct ev_signal *w) +#if EV_EMBED_ENABLE +void noinline +ev_embed_sweep (EV_P_ ev_embed *w) { - if (ev_is_active (w)) - return; + ev_loop (w->other, EVLOOP_NONBLOCK); +} - ev_start ((W)w, 1); - array_needsize (signals, signalmax, w->signum, signals_init); - wlist_add ((WL *)&signals [w->signum - 1].head, (WL)w); +static void +embed_io_cb (EV_P_ ev_io *io, int revents) +{ + ev_embed *w = (ev_embed *)(((char *)io) - offsetof (ev_embed, io)); - if (!w->next) - { - struct sigaction sa; - sa.sa_handler = sighandler; - sigfillset (&sa.sa_mask); - sa.sa_flags = 0; - sigaction (w->signum, &sa, 0); - } + if (ev_cb (w)) + ev_feed_event (EV_A_ (W)w, EV_EMBED); + else + ev_loop (w->other, EVLOOP_NONBLOCK); } -void -evsignal_stop (struct ev_signal *w) +static void +embed_prepare_cb (EV_P_ ev_prepare *prepare, int revents) { - if (!ev_is_active (w)) - return; + ev_embed *w = (ev_embed *)(((char *)prepare) - offsetof (ev_embed, prepare)); - wlist_del ((WL *)&signals [w->signum - 1].head, (WL)w); - ev_stop ((W)w); + { + struct ev_loop *loop = w->other; - if (!signals [w->signum - 1].head) - signal (w->signum, SIG_DFL); + while (fdchangecnt) + { + fd_reify (EV_A); + ev_loop (EV_A_ EVLOOP_NONBLOCK); + } + } } -void evidle_start (struct ev_idle *w) +#if 0 +static void +embed_idle_cb (EV_P_ ev_idle *idle, int revents) { - if (ev_is_active (w)) + ev_idle_stop (EV_A_ idle); +} +#endif + +void +ev_embed_start (EV_P_ ev_embed *w) +{ + if (expect_false (ev_is_active (w))) return; - ev_start ((W)w, ++idlecnt); - array_needsize (idles, idlemax, idlecnt, ); - idles [idlecnt - 1] = w; + { + struct ev_loop *loop = w->other; + assert (("loop to be embedded is not embeddable", backend & ev_embeddable_backends ())); + ev_io_init (&w->io, embed_io_cb, backend_fd, EV_READ); + } + + ev_set_priority (&w->io, ev_priority (w)); + ev_io_start (EV_A_ &w->io); + + ev_prepare_init (&w->prepare, embed_prepare_cb); + ev_set_priority (&w->prepare, EV_MINPRI); + ev_prepare_start (EV_A_ &w->prepare); + + /*ev_idle_init (&w->idle, e,bed_idle_cb);*/ + + ev_start (EV_A_ (W)w, 1); } -void evidle_stop (struct ev_idle *w) +void +ev_embed_stop (EV_P_ ev_embed *w) { - idles [w->active - 1] = idles [--idlecnt]; - ev_stop ((W)w); + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; + + ev_io_stop (EV_A_ &w->io); + ev_prepare_stop (EV_A_ &w->prepare); + + ev_stop (EV_A_ (W)w); } +#endif -void evcheck_start (struct ev_check *w) +#if EV_FORK_ENABLE +void +ev_fork_start (EV_P_ ev_fork *w) { - if (ev_is_active (w)) + if (expect_false (ev_is_active (w))) return; - ev_start ((W)w, ++checkcnt); - array_needsize (checks, checkmax, checkcnt, ); - checks [checkcnt - 1] = w; + ev_start (EV_A_ (W)w, ++forkcnt); + array_needsize (ev_fork *, forks, forkmax, forkcnt, EMPTY2); + forks [forkcnt - 1] = w; } -void evcheck_stop (struct ev_check *w) +void +ev_fork_stop (EV_P_ ev_fork *w) { - checks [w->active - 1] = checks [--checkcnt]; - ev_stop ((W)w); -} + clear_pending (EV_A_ (W)w); + if (expect_false (!ev_is_active (w))) + return; -/*****************************************************************************/ + { + int active = ((W)w)->active; + forks [active - 1] = forks [--forkcnt]; + ((W)forks [active - 1])->active = active; + } -#if 1 + ev_stop (EV_A_ (W)w); +} +#endif -struct ev_io wio; +/*****************************************************************************/ -static void -sin_cb (struct ev_io *w, int revents) +struct ev_once { - fprintf (stderr, "sin %d, revents %d\n", w->fd, revents); -} + ev_io io; + ev_timer to; + void (*cb)(int revents, void *arg); + void *arg; +}; static void -ocb (struct ev_timer *w, int revents) +once_cb (EV_P_ struct ev_once *once, int revents) { - //fprintf (stderr, "timer %f,%f (%x) (%f) d%p\n", w->at, w->repeat, revents, w->at - ev_time (), w->data); - evtimer_stop (w); - evtimer_start (w); + void (*cb)(int revents, void *arg) = once->cb; + void *arg = once->arg; + + ev_io_stop (EV_A_ &once->io); + ev_timer_stop (EV_A_ &once->to); + ev_free (once); + + cb (revents, arg); } static void -scb (struct ev_signal *w, int revents) +once_cb_io (EV_P_ ev_io *w, int revents) { - fprintf (stderr, "signal %x,%d\n", revents, w->signum); - evio_stop (&wio); - evio_start (&wio); + once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, io)), revents); } static void -gcb (struct ev_signal *w, int revents) +once_cb_to (EV_P_ ev_timer *w, int revents) { - fprintf (stderr, "generic %x\n", revents); - + once_cb (EV_A_ (struct ev_once *)(((char *)w) - offsetof (struct ev_once, to)), revents); } -int main (void) +void +ev_once (EV_P_ int fd, int events, ev_tstamp timeout, void (*cb)(int revents, void *arg), void *arg) { - ev_init (0); - - evio_init (&wio, sin_cb, 0, EV_READ); - evio_start (&wio); + struct ev_once *once = (struct ev_once *)ev_malloc (sizeof (struct ev_once)); - struct ev_timer t[10000]; - -#if 0 - int i; - for (i = 0; i < 10000; ++i) + if (expect_false (!once)) { - struct ev_timer *w = t + i; - evw_init (w, ocb, i); - evtimer_init_abs (w, ocb, drand48 (), 0.99775533); - evtimer_start (w); - if (drand48 () < 0.5) - evtimer_stop (w); + cb (EV_ERROR | EV_READ | EV_WRITE | EV_TIMEOUT, arg); + return; } -#endif - struct ev_timer t1; - evtimer_init (&t1, ocb, 5, 10); - evtimer_start (&t1); + once->cb = cb; + once->arg = arg; - struct ev_signal sig; - evsignal_init (&sig, scb, SIGQUIT); - evsignal_start (&sig); - - struct ev_check cw; - evcheck_init (&cw, gcb); - evcheck_start (&cw); - - struct ev_idle iw; - evidle_init (&iw, gcb); - evidle_start (&iw); - - ev_loop (0); + ev_init (&once->io, once_cb_io); + if (fd >= 0) + { + ev_io_set (&once->io, fd, events); + ev_io_start (EV_A_ &once->io); + } - return 0; + ev_init (&once->to, once_cb_to); + if (timeout >= 0.) + { + ev_timer_set (&once->to, timeout, 0.); + ev_timer_start (EV_A_ &once->to); + } } +#if EV_MULTIPLICITY + #include "ev_wrap.h" #endif - - +#ifdef __cplusplus +} +#endif