* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
#ifndef EV_STANDALONE
# include "config.h"
# define EV_USE_EPOLL 1
# endif
-# if HAVE_KQUEUE && HAVE_WORKING_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H
+# if HAVE_KQUEUE && HAVE_SYS_EVENT_H && HAVE_SYS_QUEUE_H
# define EV_USE_KQUEUE 1
# endif
#define PID_HASHSIZE 16 /* size of pid hash table, must be power of two */
/*#define CLEANUP_INTERVAL 300. /* how often to try to free memory and re-check fds */
-#include "ev.h"
+#ifdef EV_H
+# include EV_H
+#else
+# include "ev.h"
+#endif
#if __GNUC__ >= 3
# define expect(expr,value) __builtin_expect ((expr),(value))
#if EV_MULTIPLICITY
-struct ev_loop
-{
-# define VAR(name,decl) decl;
-# include "ev_vars.h"
-};
-# undef VAR
-# include "ev_wrap.h"
+ struct ev_loop
+ {
+ ev_tstamp ev_rt_now;
+ #define VAR(name,decl) decl;
+ #include "ev_vars.h"
+ #undef VAR
+ };
+ #include "ev_wrap.h"
+
+ struct ev_loop default_loop_struct;
+ static struct ev_loop *default_loop;
#else
-# define VAR(name,decl) static decl;
-# include "ev_vars.h"
-# undef VAR
+ ev_tstamp ev_rt_now;
+ #define VAR(name,decl) static decl;
+ #include "ev_vars.h"
+ #undef VAR
+
+ static int default_loop;
#endif
return ev_time ();
}
+#if EV_MULTIPLICITY
ev_tstamp
ev_now (EV_P)
{
- return rt_now;
+ return ev_rt_now;
}
+#endif
#define array_roundsize(type,n) ((n) | 4 & ~3)
}
}
-static void
-event (EV_P_ W w, int events)
+void
+ev_feed_event (EV_P_ void *w, int revents)
{
- if (w->pending)
+ W w_ = (W)w;
+
+ if (w_->pending)
{
- pendings [ABSPRI (w)][w->pending - 1].events |= events;
+ pendings [ABSPRI (w_)][w_->pending - 1].events |= revents;
return;
}
- w->pending = ++pendingcnt [ABSPRI (w)];
- array_needsize (ANPENDING, pendings [ABSPRI (w)], pendingmax [ABSPRI (w)], pendingcnt [ABSPRI (w)], (void));
- pendings [ABSPRI (w)][w->pending - 1].w = w;
- pendings [ABSPRI (w)][w->pending - 1].events = events;
+ w_->pending = ++pendingcnt [ABSPRI (w_)];
+ array_needsize (ANPENDING, pendings [ABSPRI (w_)], pendingmax [ABSPRI (w_)], pendingcnt [ABSPRI (w_)], (void));
+ pendings [ABSPRI (w_)][w_->pending - 1].w = w_;
+ pendings [ABSPRI (w_)][w_->pending - 1].events = revents;
}
static void
int i;
for (i = 0; i < eventcnt; ++i)
- event (EV_A_ events [i], type);
+ ev_feed_event (EV_A_ events [i], type);
}
-static void
-fd_event (EV_P_ int fd, int events)
+inline void
+fd_event (EV_P_ int fd, int revents)
{
ANFD *anfd = anfds + fd;
struct ev_io *w;
for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next)
{
- int ev = w->events & events;
+ int ev = w->events & revents;
if (ev)
- event (EV_A_ (W)w, ev);
+ ev_feed_event (EV_A_ (W)w, ev);
}
}
+void
+ev_feed_fd_event (EV_P_ int fd, int revents)
+{
+ fd_event (EV_A_ fd, revents);
+}
+
/*****************************************************************************/
static void
while ((w = (struct ev_io *)anfds [fd].head))
{
ev_io_stop (EV_A_ w);
- event (EV_A_ (W)w, EV_ERROR | EV_READ | EV_WRITE);
+ ev_feed_event (EV_A_ (W)w, EV_ERROR | EV_READ | EV_WRITE);
}
}
((W)heap [k])->active = k + 1;
}
+inline void
+adjustheap (WT *heap, int N, int k, ev_tstamp at)
+{
+ ev_tstamp old_at = heap [k]->at;
+ heap [k]->at = at;
+
+ if (old_at < at)
+ downheap (heap, N, k);
+ else
+ upheap (heap, k);
+}
+
/*****************************************************************************/
typedef struct
}
}
+void
+ev_feed_signal_event (EV_P_ int signum)
+{
+ WL w;
+
+#if EV_MULTIPLICITY
+ assert (("feeding signal events is only supported in the default loop", loop == default_loop));
+#endif
+
+ --signum;
+
+ if (signum < 0 || signum >= signalmax)
+ return;
+
+ signals [signum].gotsig = 0;
+
+ for (w = signals [signum].head; w; w = w->next)
+ ev_feed_event (EV_A_ (W)w, EV_SIGNAL);
+}
+
static void
sigcb (EV_P_ struct ev_io *iow, int revents)
{
- WL w;
int signum;
#ifdef WIN32
for (signum = signalmax; signum--; )
if (signals [signum].gotsig)
- {
- signals [signum].gotsig = 0;
-
- for (w = signals [signum].head; w; w = w->next)
- event (EV_A_ (W)w, EV_SIGNAL);
- }
+ ev_feed_signal_event (EV_A_ signum + 1);
}
static void
ev_priority (w) = ev_priority (sw); /* need to do it *now* */
w->rpid = pid;
w->rstatus = status;
- event (EV_A_ (W)w, EV_CHILD);
+ ev_feed_event (EV_A_ (W)w, EV_CHILD);
}
}
if (0 < (pid = waitpid (-1, &status, WNOHANG | WUNTRACED | WCONTINUED)))
{
/* make sure we are called again until all childs have been reaped */
- event (EV_A_ (W)sw, EV_SIGNAL);
+ ev_feed_event (EV_A_ (W)sw, EV_SIGNAL);
child_reap (EV_A_ sw, pid, pid, status);
child_reap (EV_A_ sw, 0, pid, status); /* this might trigger a watcher twice, but event catches that */
}
#endif
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
mn_now = get_clock ();
now_floor = mn_now;
- rtmn_diff = rt_now - mn_now;
+ rtmn_diff = ev_rt_now - mn_now;
if (methods == EVMETHOD_AUTO)
if (!enable_secure () && getenv ("LIBEV_METHODS"))
if (!method && (methods & EVMETHOD_SELECT)) method = select_init (EV_A_ methods);
#endif
- ev_watcher_init (&sigev, sigcb);
+ ev_init (&sigev, sigcb);
ev_set_priority (&sigev, EV_MAXPRI);
}
}
#endif
#if EV_MULTIPLICITY
-struct ev_loop default_loop_struct;
-static struct ev_loop *default_loop;
-
struct ev_loop *
#else
-static int default_loop;
-
int
#endif
ev_default_loop (int methods)
if (p->w)
{
p->w->pending = 0;
- p->w->cb (EV_A_ p->w, p->events);
+ EV_CB_INVOKE (p->w, p->events);
}
}
}
else
ev_timer_stop (EV_A_ w); /* nonrepeating: stop timer */
- event (EV_A_ (W)w, EV_TIMEOUT);
+ ev_feed_event (EV_A_ (W)w, EV_TIMEOUT);
}
}
static void
periodics_reify (EV_P)
{
- while (periodiccnt && ((WT)periodics [0])->at <= rt_now)
+ while (periodiccnt && ((WT)periodics [0])->at <= ev_rt_now)
{
struct ev_periodic *w = periodics [0];
assert (("inactive timer on periodic heap detected", ev_is_active (w)));
/* first reschedule or stop timer */
- if (w->interval)
+ if (w->reschedule_cb)
{
- ((WT)w)->at += floor ((rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval;
- assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > rt_now));
+ ev_tstamp at = ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + 0.0001);
+
+ assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > ev_rt_now));
+ downheap ((WT *)periodics, periodiccnt, 0);
+ }
+ else if (w->interval)
+ {
+ ((WT)w)->at += floor ((ev_rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval;
+ assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > ev_rt_now));
downheap ((WT *)periodics, periodiccnt, 0);
}
else
ev_periodic_stop (EV_A_ w); /* nonrepeating: stop timer */
- event (EV_A_ (W)w, EV_PERIODIC);
+ ev_feed_event (EV_A_ (W)w, EV_PERIODIC);
}
}
{
struct ev_periodic *w = periodics [i];
- if (w->interval)
- {
- ev_tstamp diff = ceil ((rt_now - ((WT)w)->at) / w->interval) * w->interval;
-
- if (fabs (diff) >= 1e-4)
- {
- ev_periodic_stop (EV_A_ w);
- ev_periodic_start (EV_A_ w);
-
- i = 0; /* restart loop, inefficient, but time jumps should be rare */
- }
- }
+ if (w->reschedule_cb)
+ ((WT)w)->at = w->reschedule_cb (w, ev_rt_now);
+ else if (w->interval)
+ ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval;
}
+
+ /* now rebuild the heap */
+ for (i = periodiccnt >> 1; i--; )
+ downheap ((WT *)periodics, periodiccnt, i);
}
inline int
if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5))
{
- rt_now = rtmn_diff + mn_now;
+ ev_rt_now = rtmn_diff + mn_now;
return 0;
}
else
{
now_floor = mn_now;
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
return 1;
}
}
for (i = 4; --i; ) /* loop a few times, before making important decisions */
{
- rtmn_diff = rt_now - mn_now;
+ rtmn_diff = ev_rt_now - mn_now;
if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP)
return; /* all is well */
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
mn_now = get_clock ();
now_floor = mn_now;
}
else
#endif
{
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
- if (expect_false (mn_now > rt_now || mn_now < rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP))
+ if (expect_false (mn_now > ev_rt_now || mn_now < ev_rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP))
{
periodics_reschedule (EV_A);
/* adjust timers. this is easy, as the offset is the same for all */
for (i = 0; i < timercnt; ++i)
- ((WT)timers [i])->at += rt_now - mn_now;
+ ((WT)timers [i])->at += ev_rt_now - mn_now;
}
- mn_now = rt_now;
+ mn_now = ev_rt_now;
}
}
else
#endif
{
- rt_now = ev_time ();
- mn_now = rt_now;
+ ev_rt_now = ev_time ();
+ mn_now = ev_rt_now;
}
if (flags & EVLOOP_NONBLOCK || idlecnt)
if (periodiccnt)
{
- ev_tstamp to = ((WT)periodics [0])->at - rt_now + method_fudge;
+ ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + method_fudge;
if (block > to) block = to;
}
method_poll (EV_A_ block);
- /* update rt_now, do magic */
+ /* update ev_rt_now, do magic */
time_update (EV_A);
/* queue pending timers and reschedule them */
if (ev_is_active (w))
{
if (w->repeat)
- {
- ((WT)w)->at = mn_now + w->repeat;
- downheap ((WT *)timers, timercnt, ((W)w)->active - 1);
- }
+ adjustheap ((WT *)timers, timercnt, ((W)w)->active - 1, mn_now + w->repeat);
else
ev_timer_stop (EV_A_ w);
}
if (ev_is_active (w))
return;
- assert (("ev_periodic_start called with negative interval value", w->interval >= 0.));
-
- /* this formula differs from the one in periodic_reify because we do not always round up */
- if (w->interval)
- ((WT)w)->at += ceil ((rt_now - ((WT)w)->at) / w->interval) * w->interval;
+ if (w->reschedule_cb)
+ ((WT)w)->at = w->reschedule_cb (w, ev_rt_now);
+ else if (w->interval)
+ {
+ assert (("ev_periodic_start called with negative interval value", w->interval >= 0.));
+ /* this formula differs from the one in periodic_reify because we do not always round up */
+ ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval;
+ }
ev_start (EV_A_ (W)w, ++periodiccnt);
array_needsize (struct ev_periodic *, periodics, periodicmax, periodiccnt, (void));
ev_stop (EV_A_ (W)w);
}
+void
+ev_periodic_again (EV_P_ struct ev_periodic *w)
+{
+ /* TODO: use adjustheap and recalculation */
+ ev_periodic_stop (EV_A_ w);
+ ev_periodic_start (EV_A_ w);
+}
+
void
ev_idle_start (EV_P_ struct ev_idle *w)
{
once->cb = cb;
once->arg = arg;
- ev_watcher_init (&once->io, once_cb_io);
+ ev_init (&once->io, once_cb_io);
if (fd >= 0)
{
ev_io_set (&once->io, fd, events);
ev_io_start (EV_A_ &once->io);
}
- ev_watcher_init (&once->to, once_cb_to);
+ ev_init (&once->to, once_cb_to);
if (timeout >= 0.)
{
ev_timer_set (&once->to, timeout, 0.);
}
}
+#ifdef __cplusplus
+}
+#endif
+