X-Git-Url: https://git.llucax.com/software/libev.git/blobdiff_plain/e53306e99926168d590f09e39ec05ad229799811..4ce77a5b1de05dae6092eae478b2a4397173d9c3:/ev.c?ds=sidebyside diff --git a/ev.c b/ev.c index f8b08f5..acde504 100644 --- a/ev.c +++ b/ev.c @@ -218,34 +218,45 @@ extern "C" { /**/ +/* + * This is used to avoid floating point rounding problems. + * It is added to ev_rt_now when scheduling periodics + * to ensure progress, time-wise, even when rounding + * errors are against us. + * This value is good at least till the year 4000. + * Better solutions welcome. + */ +#define TIME_EPSILON 0.0001220703125 /* 1/8192 */ + #define MIN_TIMEJUMP 1. /* minimum timejump that gets detected (if monotonic clock available) */ #define MAX_BLOCKTIME 59.743 /* never wait longer than this time (to detect time jumps) */ -/*#define CLEANUP_INTERVAL (MAX_BLOCKTIME * 5.) /* how often to try to free memory and re-check fds */ +/*#define CLEANUP_INTERVAL (MAX_BLOCKTIME * 5.) /* how often to try to free memory and re-check fds, TODO */ #if __GNUC__ >= 3 # define expect(expr,value) __builtin_expect ((expr),(value)) -# define inline_size static inline /* inline for codesize */ -# if EV_MINIMAL -# define noinline __attribute__ ((noinline)) -# define inline_speed static noinline -# else -# define noinline -# define inline_speed static inline -# endif +# define noinline __attribute__ ((noinline)) #else # define expect(expr,value) (expr) -# define inline_speed static -# define inline_size static # define noinline +# if __STDC_VERSION__ < 199901L +# define inline +# endif #endif #define expect_false(expr) expect ((expr) != 0, 0) #define expect_true(expr) expect ((expr) != 0, 1) +#define inline_size static inline + +#if EV_MINIMAL +# define inline_speed static noinline +#else +# define inline_speed static inline +#endif #define NUMPRI (EV_MAXPRI - EV_MINPRI + 1) -#define ABSPRI(w) ((w)->priority - EV_MINPRI) +#define ABSPRI(w) (((W)w)->priority - EV_MINPRI) -#define EMPTY0 /* required for microsofts broken pseudo-c compiler */ +#define EMPTY /* required for microsofts broken pseudo-c compiler */ #define EMPTY2(a,b) /* used to suppress some warnings */ typedef ev_watcher *W; @@ -283,22 +294,22 @@ syserr (const char *msg) } } -static void *(*alloc)(void *ptr, size_t size) = realloc; +static void *(*alloc)(void *ptr, long size); void -ev_set_allocator (void *(*cb)(void *ptr, size_t size)) +ev_set_allocator (void *(*cb)(void *ptr, long size)) { alloc = cb; } inline_speed void * -ev_realloc (void *ptr, size_t size) +ev_realloc (void *ptr, long size) { - ptr = alloc (ptr, size); + ptr = alloc ? alloc (ptr, size) : realloc (ptr, size); if (!ptr && size) { - fprintf (stderr, "libev: cannot allocate %ld bytes, aborting.", (long)size); + fprintf (stderr, "libev: cannot allocate %ld bytes, aborting.", size); abort (); } @@ -326,12 +337,12 @@ typedef struct int events; } ANPENDING; +#if EV_USE_INOTIFY typedef struct { -#if EV_USE_INOTIFY WL head; -#endif } ANFS; +#endif #if EV_MULTIPLICITY @@ -398,23 +409,44 @@ ev_now (EV_P) } #endif -#define array_roundsize(type,n) (((n) | 4) & ~3) +int inline_size +array_nextsize (int elem, int cur, int cnt) +{ + int ncur = cur + 1; + + do + ncur <<= 1; + while (cnt > ncur); + + /* if size > 4096, round to 4096 - 4 * longs to accomodate malloc overhead */ + if (elem * ncur > 4096) + { + ncur *= elem; + ncur = (ncur + elem + 4095 + sizeof (void *) * 4) & ~4095; + ncur = ncur - sizeof (void *) * 4; + ncur /= elem; + } + + return ncur; +} + +static noinline void * +array_realloc (int elem, void *base, int *cur, int cnt) +{ + *cur = array_nextsize (elem, *cur, cnt); + return ev_realloc (base, elem * *cur); +} #define array_needsize(type,base,cur,cnt,init) \ - if (expect_false ((cnt) > cur)) \ + if (expect_false ((cnt) > (cur))) \ { \ - int newcnt = cur; \ - do \ - { \ - newcnt = array_roundsize (type, newcnt << 1); \ - } \ - while ((cnt) > newcnt); \ - \ - base = (type *)ev_realloc (base, sizeof (type) * (newcnt));\ - init (base + cur, newcnt - cur); \ - cur = newcnt; \ + int ocur_ = (cur); \ + (base) = (type *)array_realloc \ + (sizeof (type), (base), &(cur), (cnt)); \ + init ((base) + (ocur_), (cur) - ocur_); \ } +#if 0 #define array_slim(type,stem) \ if (stem ## max < array_roundsize (stem ## cnt >> 2)) \ { \ @@ -422,6 +454,7 @@ ev_now (EV_P) base = (type *)ev_realloc (base, sizeof (type) * (stem ## max));\ fprintf (stderr, "slimmed down " # stem " to %d\n", stem ## max);/*D*/\ } +#endif #define array_free(stem, idx) \ ev_free (stem ## s idx); stem ## cnt idx = stem ## max idx = 0; @@ -432,17 +465,17 @@ void noinline ev_feed_event (EV_P_ void *w, int revents) { W w_ = (W)w; + int pri = ABSPRI (w_); if (expect_false (w_->pending)) + pendings [pri][w_->pending - 1].events |= revents; + else { - pendings [ABSPRI (w_)][w_->pending - 1].events |= revents; - return; + w_->pending = ++pendingcnt [pri]; + array_needsize (ANPENDING, pendings [pri], pendingmax [pri], w_->pending, EMPTY2); + pendings [pri][w_->pending - 1].w = w_; + pendings [pri][w_->pending - 1].events = revents; } - - w_->pending = ++pendingcnt [ABSPRI (w_)]; - array_needsize (ANPENDING, pendings [ABSPRI (w_)], pendingmax [ABSPRI (w_)], pendingcnt [ABSPRI (w_)], EMPTY2); - pendings [ABSPRI (w_)][w_->pending - 1].w = w_; - pendings [ABSPRI (w_)][w_->pending - 1].events = revents; } void inline_size @@ -487,7 +520,8 @@ fd_event (EV_P_ int fd, int revents) void ev_feed_fd_event (EV_P_ int fd, int revents) { - fd_event (EV_A_ fd, revents); + if (fd >= 0 && fd < anfdmax) + fd_event (EV_A_ fd, revents); } void inline_size @@ -591,7 +625,6 @@ fd_rearm_all (EV_P) { int fd; - /* this should be highly optimised to not do anything but set a flag */ for (fd = 0; fd < anfdmax; ++fd) if (anfds [fd].events) { @@ -728,7 +761,7 @@ sigcb (EV_P_ ev_io *iow, int revents) ev_feed_signal_event (EV_A_ signum + 1); } -void inline_size +void inline_speed fd_intern (int fd) { #ifdef _WIN32 @@ -767,9 +800,9 @@ child_reap (EV_P_ ev_signal *sw, int chain, int pid, int status) for (w = (ev_child *)childs [chain & (EV_PID_HASHSIZE - 1)]; w; w = (ev_child *)((WL)w)->next) if (w->pid == pid || !w->pid) { - ev_priority (w) = ev_priority (sw); /* need to do it *now* */ - w->rpid = pid; - w->rstatus = status; + ev_set_priority (w, ev_priority (sw)); /* need to do it *now* */ + w->rpid = pid; + w->rstatus = status; ev_feed_event (EV_A_ (W)w, EV_CHILD); } } @@ -889,6 +922,12 @@ ev_backend (EV_P) return backend; } +unsigned int +ev_loop_count (EV_P) +{ + return loop_count; +} + static void noinline loop_init (EV_P_ unsigned int flags) { @@ -907,6 +946,12 @@ loop_init (EV_P_ unsigned int flags) now_floor = mn_now; rtmn_diff = ev_rt_now - mn_now; + /* pid check not overridable via env */ +#ifndef _WIN32 + if (flags & EVFLAG_FORKCHECK) + curpid = getpid (); +#endif + if (!(flags & EVFLAG_NOENV) && !enable_secure () && getenv ("LIBEV_FLAGS")) @@ -972,17 +1017,21 @@ loop_destroy (EV_P) #endif for (i = NUMPRI; i--; ) - array_free (pending, [i]); + { + array_free (pending, [i]); +#if EV_IDLE_ENABLE + array_free (idle, [i]); +#endif + } /* have to use the microsoft-never-gets-it-right macro */ - array_free (fdchange, EMPTY0); - array_free (timer, EMPTY0); + array_free (fdchange, EMPTY); + array_free (timer, EMPTY); #if EV_PERIODIC_ENABLE - array_free (periodic, EMPTY0); + array_free (periodic, EMPTY); #endif - array_free (idle, EMPTY0); - array_free (prepare, EMPTY0); - array_free (check, EMPTY0); + array_free (prepare, EMPTY); + array_free (check, EMPTY); backend = 0; } @@ -1128,16 +1177,10 @@ ev_default_fork (void) /*****************************************************************************/ -int inline_size -any_pending (EV_P) +void +ev_invoke (EV_P_ void *w, int revents) { - int pri; - - for (pri = NUMPRI; pri--; ) - if (pendingcnt [pri]) - return 1; - - return 0; + EV_CB_INVOKE ((W)w, revents); } void inline_speed @@ -1200,13 +1243,14 @@ periodics_reify (EV_P) /* first reschedule or stop timer */ if (w->reschedule_cb) { - ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + 0.0001); + ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + TIME_EPSILON); assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > ev_rt_now)); downheap ((WT *)periodics, periodiccnt, 0); } else if (w->interval) { - ((WT)w)->at += floor ((ev_rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval; + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; + if (((WT)w)->at - ev_rt_now <= TIME_EPSILON) ((WT)w)->at += w->interval; assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > ev_rt_now)); downheap ((WT *)periodics, periodiccnt, 0); } @@ -1230,7 +1274,7 @@ periodics_reschedule (EV_P) if (w->reschedule_cb) ((WT)w)->at = w->reschedule_cb (w, ev_rt_now); else if (w->interval) - ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval; + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; } /* now rebuild the heap */ @@ -1239,75 +1283,89 @@ periodics_reschedule (EV_P) } #endif -int inline_size -time_update_monotonic (EV_P) +#if EV_IDLE_ENABLE +void inline_size +idle_reify (EV_P) { - mn_now = get_clock (); - - if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5)) + if (expect_false (idleall)) { - ev_rt_now = rtmn_diff + mn_now; - return 0; - } - else - { - now_floor = mn_now; - ev_rt_now = ev_time (); - return 1; + int pri; + + for (pri = NUMPRI; pri--; ) + { + if (pendingcnt [pri]) + break; + + if (idlecnt [pri]) + { + queue_events (EV_A_ (W *)idles [pri], idlecnt [pri], EV_IDLE); + break; + } + } } } +#endif -void inline_size -time_update (EV_P) +void inline_speed +time_update (EV_P_ ev_tstamp max_block) { int i; #if EV_USE_MONOTONIC if (expect_true (have_monotonic)) { - if (time_update_monotonic (EV_A)) + ev_tstamp odiff = rtmn_diff; + + mn_now = get_clock (); + + /* only fetch the realtime clock every 0.5*MIN_TIMEJUMP seconds */ + /* interpolate in the meantime */ + if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5)) { - ev_tstamp odiff = rtmn_diff; - - /* loop a few times, before making important decisions. - * on the choice of "4": one iteration isn't enough, - * in case we get preempted during the calls to - * ev_time and get_clock. a second call is almost guarenteed - * to succeed in that case, though. and looping a few more times - * doesn't hurt either as we only do this on time-jumps or - * in the unlikely event of getting preempted here. - */ - for (i = 4; --i; ) - { - rtmn_diff = ev_rt_now - mn_now; + ev_rt_now = rtmn_diff + mn_now; + return; + } - if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP) - return; /* all is well */ + now_floor = mn_now; + ev_rt_now = ev_time (); - ev_rt_now = ev_time (); - mn_now = get_clock (); - now_floor = mn_now; - } + /* loop a few times, before making important decisions. + * on the choice of "4": one iteration isn't enough, + * in case we get preempted during the calls to + * ev_time and get_clock. a second call is almost guaranteed + * to succeed in that case, though. and looping a few more times + * doesn't hurt either as we only do this on time-jumps or + * in the unlikely event of having been preempted here. + */ + for (i = 4; --i; ) + { + rtmn_diff = ev_rt_now - mn_now; + + if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP) + return; /* all is well */ + + ev_rt_now = ev_time (); + mn_now = get_clock (); + now_floor = mn_now; + } # if EV_PERIODIC_ENABLE - periodics_reschedule (EV_A); + periodics_reschedule (EV_A); # endif - /* no timer adjustment, as the monotonic clock doesn't jump */ - /* timers_reschedule (EV_A_ rtmn_diff - odiff) */ - } + /* no timer adjustment, as the monotonic clock doesn't jump */ + /* timers_reschedule (EV_A_ rtmn_diff - odiff) */ } else #endif { ev_rt_now = ev_time (); - if (expect_false (mn_now > ev_rt_now || mn_now < ev_rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP)) + if (expect_false (mn_now > ev_rt_now || ev_rt_now > mn_now + max_block + MIN_TIMEJUMP)) { #if EV_PERIODIC_ENABLE periodics_reschedule (EV_A); #endif - - /* adjust timers. this is easy, as the offset is the same for all */ + /* adjust timers. this is easy, as the offset is the same for all of them */ for (i = 0; i < timercnt; ++i) ((WT)timers [i])->at += ev_rt_now - mn_now; } @@ -1337,25 +1395,39 @@ ev_loop (EV_P_ int flags) ? EVUNLOOP_ONE : EVUNLOOP_CANCEL; - while (activecnt) + call_pending (EV_A); /* in case we recurse, ensure ordering stays nice and clean */ + + do { - /* we might have forked, so reify kernel state if necessary */ - #if EV_FORK_ENABLE - if (expect_false (postfork)) - if (forkcnt) - { - queue_events (EV_A_ (W *)forks, forkcnt, EV_FORK); - call_pending (EV_A); - } - #endif +#ifndef _WIN32 + if (expect_false (curpid)) /* penalise the forking check even more */ + if (expect_false (getpid () != curpid)) + { + curpid = getpid (); + postfork = 1; + } +#endif + +#if EV_FORK_ENABLE + /* we might have forked, so queue fork handlers */ + if (expect_false (postfork)) + if (forkcnt) + { + queue_events (EV_A_ (W *)forks, forkcnt, EV_FORK); + call_pending (EV_A); + } +#endif - /* queue check watchers (and execute them) */ + /* queue prepare watchers (and execute them) */ if (expect_false (preparecnt)) { queue_events (EV_A_ (W *)prepares, preparecnt, EV_PREPARE); call_pending (EV_A); } + if (expect_false (!activecnt)) + break; + /* we might have forked, so reify kernel state if necessary */ if (expect_false (postfork)) loop_fork (EV_A); @@ -1365,22 +1437,14 @@ ev_loop (EV_P_ int flags) /* calculate blocking time */ { - double block; + ev_tstamp block; - if (flags & EVLOOP_NONBLOCK || idlecnt) + if (expect_false (flags & EVLOOP_NONBLOCK || idleall || !activecnt)) block = 0.; /* do not block at all */ else { /* update time to cancel out callback processing overhead */ -#if EV_USE_MONOTONIC - if (expect_true (have_monotonic)) - time_update_monotonic (EV_A); - else -#endif - { - ev_rt_now = ev_time (); - mn_now = ev_rt_now; - } + time_update (EV_A_ 1e100); block = MAX_BLOCKTIME; @@ -1401,11 +1465,12 @@ ev_loop (EV_P_ int flags) if (expect_false (block < 0.)) block = 0.; } + ++loop_count; backend_poll (EV_A_ block); - } - /* update ev_rt_now, do magic */ - time_update (EV_A); + /* update ev_rt_now, do magic */ + time_update (EV_A_ block); + } /* queue pending timers and reschedule them */ timers_reify (EV_A); /* relative timers called last */ @@ -1413,9 +1478,10 @@ ev_loop (EV_P_ int flags) periodics_reify (EV_A); /* absolute timers called first */ #endif +#if EV_IDLE_ENABLE /* queue idle watchers unless other events are pending */ - if (idlecnt && !any_pending (EV_A)) - queue_events (EV_A_ (W *)idles, idlecnt, EV_IDLE); + idle_reify (EV_A); +#endif /* queue check watchers, to be executed first */ if (expect_false (checkcnt)) @@ -1423,9 +1489,8 @@ ev_loop (EV_P_ int flags) call_pending (EV_A); - if (expect_false (loop_done)) - break; } + while (expect_true (activecnt && !loop_done)); if (loop_done == EVUNLOOP_ONE) loop_done = EVUNLOOP_CANCEL; @@ -1462,7 +1527,7 @@ wlist_del (WL *head, WL elem) } void inline_speed -ev_clear_pending (EV_P_ W w) +clear_pending (EV_P_ W w) { if (w->pending) { @@ -1471,12 +1536,36 @@ ev_clear_pending (EV_P_ W w) } } +int +ev_clear_pending (EV_P_ void *w) +{ + W w_ = (W)w; + int pending = w_->pending; + + if (expect_true (pending)) + { + ANPENDING *p = pendings [ABSPRI (w_)] + pending - 1; + w_->pending = 0; + p->w = 0; + return p->events; + } + else + return 0; +} + +void inline_size +pri_adjust (EV_P_ W w) +{ + int pri = w->priority; + pri = pri < EV_MINPRI ? EV_MINPRI : pri; + pri = pri > EV_MAXPRI ? EV_MAXPRI : pri; + w->priority = pri; +} + void inline_speed ev_start (EV_P_ W w, int active) { - if (w->priority < EV_MINPRI) w->priority = EV_MINPRI; - if (w->priority > EV_MAXPRI) w->priority = EV_MAXPRI; - + pri_adjust (EV_A_ w); w->active = active; ev_ref (EV_A); } @@ -1490,7 +1579,7 @@ ev_stop (EV_P_ W w) /*****************************************************************************/ -void +void noinline ev_io_start (EV_P_ ev_io *w) { int fd = w->fd; @@ -1507,10 +1596,10 @@ ev_io_start (EV_P_ ev_io *w) fd_change (EV_A_ fd); } -void +void noinline ev_io_stop (EV_P_ ev_io *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1522,7 +1611,7 @@ ev_io_stop (EV_P_ ev_io *w) fd_change (EV_A_ w->fd); } -void +void noinline ev_timer_start (EV_P_ ev_timer *w) { if (expect_false (ev_is_active (w))) @@ -1540,10 +1629,10 @@ ev_timer_start (EV_P_ ev_timer *w) /*assert (("internal timer heap corruption", timers [((W)w)->active - 1] == w));*/ } -void +void noinline ev_timer_stop (EV_P_ ev_timer *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1564,7 +1653,7 @@ ev_timer_stop (EV_P_ ev_timer *w) ev_stop (EV_A_ (W)w); } -void +void noinline ev_timer_again (EV_P_ ev_timer *w) { if (ev_is_active (w)) @@ -1585,7 +1674,7 @@ ev_timer_again (EV_P_ ev_timer *w) } #if EV_PERIODIC_ENABLE -void +void noinline ev_periodic_start (EV_P_ ev_periodic *w) { if (expect_false (ev_is_active (w))) @@ -1597,8 +1686,10 @@ ev_periodic_start (EV_P_ ev_periodic *w) { assert (("ev_periodic_start called with negative interval value", w->interval >= 0.)); /* this formula differs from the one in periodic_reify because we do not always round up */ - ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval; + ((WT)w)->at = w->offset + ceil ((ev_rt_now - w->offset) / w->interval) * w->interval; } + else + ((WT)w)->at = w->offset; ev_start (EV_A_ (W)w, ++periodiccnt); array_needsize (ev_periodic *, periodics, periodicmax, periodiccnt, EMPTY2); @@ -1608,10 +1699,10 @@ ev_periodic_start (EV_P_ ev_periodic *w) /*assert (("internal periodic heap corruption", periodics [((W)w)->active - 1] == w));*/ } -void +void noinline ev_periodic_stop (EV_P_ ev_periodic *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1630,7 +1721,7 @@ ev_periodic_stop (EV_P_ ev_periodic *w) ev_stop (EV_A_ (W)w); } -void +void noinline ev_periodic_again (EV_P_ ev_periodic *w) { /* TODO: use adjustheap and recalculation */ @@ -1643,7 +1734,7 @@ ev_periodic_again (EV_P_ ev_periodic *w) # define SA_RESTART 0 #endif -void +void noinline ev_signal_start (EV_P_ ev_signal *w) { #if EV_MULTIPLICITY @@ -1672,10 +1763,10 @@ ev_signal_start (EV_P_ ev_signal *w) } } -void +void noinline ev_signal_stop (EV_P_ ev_signal *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1702,7 +1793,7 @@ ev_child_start (EV_P_ ev_child *w) void ev_child_stop (EV_P_ ev_child *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1720,7 +1811,7 @@ ev_child_stop (EV_P_ ev_child *w) #define DEF_STAT_INTERVAL 5.0074891 #define MIN_STAT_INTERVAL 0.1074891 -void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents); +static void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents); #if EV_USE_INOTIFY # define EV_INOTIFY_BUFSIZE 8192 @@ -1881,7 +1972,7 @@ ev_stat_stat (EV_P_ ev_stat *w) w->attr.st_nlink = 1; } -void noinline +static void noinline stat_timer_cb (EV_P_ ev_timer *w_, int revents) { ev_stat *w = (ev_stat *)(((char *)w_) - offsetof (ev_stat, timer)); @@ -1891,8 +1982,20 @@ stat_timer_cb (EV_P_ ev_timer *w_, int revents) w->prev = w->attr; ev_stat_stat (EV_A_ w); - if (memcmp (&w->prev, &w->attr, sizeof (ev_statdata))) - { + /* memcmp doesn't work on netbsd, they.... do stuff to their struct stat */ + if ( + w->prev.st_dev != w->attr.st_dev + || w->prev.st_ino != w->attr.st_ino + || w->prev.st_mode != w->attr.st_mode + || w->prev.st_nlink != w->attr.st_nlink + || w->prev.st_uid != w->attr.st_uid + || w->prev.st_gid != w->attr.st_gid + || w->prev.st_rdev != w->attr.st_rdev + || w->prev.st_size != w->attr.st_size + || w->prev.st_atime != w->attr.st_atime + || w->prev.st_mtime != w->attr.st_mtime + || w->prev.st_ctime != w->attr.st_ctime + ) { #if EV_USE_INOTIFY infy_del (EV_A_ w); infy_add (EV_A_ w); @@ -1936,7 +2039,7 @@ ev_stat_start (EV_P_ ev_stat *w) void ev_stat_stop (EV_P_ ev_stat *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -1949,32 +2052,44 @@ ev_stat_stop (EV_P_ ev_stat *w) } #endif +#if EV_IDLE_ENABLE void ev_idle_start (EV_P_ ev_idle *w) { if (expect_false (ev_is_active (w))) return; - ev_start (EV_A_ (W)w, ++idlecnt); - array_needsize (ev_idle *, idles, idlemax, idlecnt, EMPTY2); - idles [idlecnt - 1] = w; + pri_adjust (EV_A_ (W)w); + + { + int active = ++idlecnt [ABSPRI (w)]; + + ++idleall; + ev_start (EV_A_ (W)w, active); + + array_needsize (ev_idle *, idles [ABSPRI (w)], idlemax [ABSPRI (w)], active, EMPTY2); + idles [ABSPRI (w)][active - 1] = w; + } } void ev_idle_stop (EV_P_ ev_idle *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; { int active = ((W)w)->active; - idles [active - 1] = idles [--idlecnt]; - ((W)idles [active - 1])->active = active; - } - ev_stop (EV_A_ (W)w); + idles [ABSPRI (w)][active - 1] = idles [ABSPRI (w)][--idlecnt [ABSPRI (w)]]; + ((W)idles [ABSPRI (w)][active - 1])->active = active; + + ev_stop (EV_A_ (W)w); + --idleall; + } } +#endif void ev_prepare_start (EV_P_ ev_prepare *w) @@ -1990,7 +2105,7 @@ ev_prepare_start (EV_P_ ev_prepare *w) void ev_prepare_stop (EV_P_ ev_prepare *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -2017,7 +2132,7 @@ ev_check_start (EV_P_ ev_check *w) void ev_check_stop (EV_P_ ev_check *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -2069,7 +2184,7 @@ ev_embed_start (EV_P_ ev_embed *w) void ev_embed_stop (EV_P_ ev_embed *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return; @@ -2094,7 +2209,7 @@ ev_fork_start (EV_P_ ev_fork *w) void ev_fork_stop (EV_P_ ev_fork *w) { - ev_clear_pending (EV_A_ (W)w); + clear_pending (EV_A_ (W)w); if (expect_false (!ev_is_active (w))) return;