* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
#ifndef EV_STANDALONE
# include "config.h"
#define PID_HASHSIZE 16 /* size of pid hash table, must be power of two */
/*#define CLEANUP_INTERVAL 300. /* how often to try to free memory and re-check fds */
-#include "ev.h"
+#ifdef EV_H
+# include EV_H
+#else
+# include "ev.h"
+#endif
#if __GNUC__ >= 3
# define expect(expr,value) __builtin_expect ((expr),(value))
#if EV_MULTIPLICITY
-struct ev_loop
-{
-# define VAR(name,decl) decl;
-# include "ev_vars.h"
-};
-# undef VAR
-# include "ev_wrap.h"
+ struct ev_loop
+ {
+ ev_tstamp ev_rt_now;
+ #define VAR(name,decl) decl;
+ #include "ev_vars.h"
+ #undef VAR
+ };
+ #include "ev_wrap.h"
+
+ struct ev_loop default_loop_struct;
+ static struct ev_loop *default_loop;
#else
-# define VAR(name,decl) static decl;
-# include "ev_vars.h"
-# undef VAR
+ ev_tstamp ev_rt_now;
+ #define VAR(name,decl) static decl;
+ #include "ev_vars.h"
+ #undef VAR
+
+ static int default_loop;
#endif
return ev_time ();
}
+#if EV_MULTIPLICITY
ev_tstamp
ev_now (EV_P)
{
- return rt_now;
+ return ev_rt_now;
}
+#endif
#define array_roundsize(type,n) ((n) | 4 & ~3)
ev_feed_event (EV_A_ events [i], type);
}
-static void
-fd_event (EV_P_ int fd, int events)
+inline void
+fd_event (EV_P_ int fd, int revents)
{
ANFD *anfd = anfds + fd;
struct ev_io *w;
for (w = (struct ev_io *)anfd->head; w; w = (struct ev_io *)((WL)w)->next)
{
- int ev = w->events & events;
+ int ev = w->events & revents;
if (ev)
ev_feed_event (EV_A_ (W)w, ev);
}
}
+void
+ev_feed_fd_event (EV_P_ int fd, int revents)
+{
+ fd_event (EV_A_ fd, revents);
+}
+
/*****************************************************************************/
static void
((W)heap [k])->active = k + 1;
}
+inline void
+adjustheap (WT *heap, int N, int k, ev_tstamp at)
+{
+ ev_tstamp old_at = heap [k]->at;
+ heap [k]->at = at;
+
+ if (old_at < at)
+ downheap (heap, N, k);
+ else
+ upheap (heap, k);
+}
+
/*****************************************************************************/
typedef struct
}
}
+void
+ev_feed_signal_event (EV_P_ int signum)
+{
+ WL w;
+
+#if EV_MULTIPLICITY
+ assert (("feeding signal events is only supported in the default loop", loop == default_loop));
+#endif
+
+ --signum;
+
+ if (signum < 0 || signum >= signalmax)
+ return;
+
+ signals [signum].gotsig = 0;
+
+ for (w = signals [signum].head; w; w = w->next)
+ ev_feed_event (EV_A_ (W)w, EV_SIGNAL);
+}
+
static void
sigcb (EV_P_ struct ev_io *iow, int revents)
{
- WL w;
int signum;
#ifdef WIN32
for (signum = signalmax; signum--; )
if (signals [signum].gotsig)
- {
- signals [signum].gotsig = 0;
-
- for (w = signals [signum].head; w; w = w->next)
- ev_feed_event (EV_A_ (W)w, EV_SIGNAL);
- }
+ ev_feed_signal_event (EV_A_ signum + 1);
}
static void
}
#endif
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
mn_now = get_clock ();
now_floor = mn_now;
- rtmn_diff = rt_now - mn_now;
+ rtmn_diff = ev_rt_now - mn_now;
if (methods == EVMETHOD_AUTO)
if (!enable_secure () && getenv ("LIBEV_METHODS"))
if (!method && (methods & EVMETHOD_SELECT)) method = select_init (EV_A_ methods);
#endif
- ev_watcher_init (&sigev, sigcb);
+ ev_init (&sigev, sigcb);
ev_set_priority (&sigev, EV_MAXPRI);
}
}
#endif
#if EV_MULTIPLICITY
-struct ev_loop default_loop_struct;
-static struct ev_loop *default_loop;
-
struct ev_loop *
#else
-static int default_loop;
-
int
#endif
ev_default_loop (int methods)
if (p->w)
{
p->w->pending = 0;
- p->w->cb (EV_A_ p->w, p->events);
+ EV_CB_INVOKE (p->w, p->events);
}
}
}
static void
periodics_reify (EV_P)
{
- while (periodiccnt && ((WT)periodics [0])->at <= rt_now)
+ while (periodiccnt && ((WT)periodics [0])->at <= ev_rt_now)
{
struct ev_periodic *w = periodics [0];
/* first reschedule or stop timer */
if (w->reschedule_cb)
{
- ev_tstamp at = ((WT)w)->at = w->reschedule_cb (w, rt_now + 0.0001);
+ ev_tstamp at = ((WT)w)->at = w->reschedule_cb (w, ev_rt_now + 0.0001);
- assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > rt_now));
+ assert (("ev_periodic reschedule callback returned time in the past", ((WT)w)->at > ev_rt_now));
downheap ((WT *)periodics, periodiccnt, 0);
}
else if (w->interval)
{
- ((WT)w)->at += floor ((rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval;
- assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > rt_now));
+ ((WT)w)->at += floor ((ev_rt_now - ((WT)w)->at) / w->interval + 1.) * w->interval;
+ assert (("ev_periodic timeout in the past detected while processing timers, negative interval?", ((WT)w)->at > ev_rt_now));
downheap ((WT *)periodics, periodiccnt, 0);
}
else
struct ev_periodic *w = periodics [i];
if (w->reschedule_cb)
- ((WT)w)->at = w->reschedule_cb (w, rt_now);
+ ((WT)w)->at = w->reschedule_cb (w, ev_rt_now);
else if (w->interval)
- ((WT)w)->at += ceil ((rt_now - ((WT)w)->at) / w->interval) * w->interval;
+ ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval;
}
/* now rebuild the heap */
if (expect_true (mn_now - now_floor < MIN_TIMEJUMP * .5))
{
- rt_now = rtmn_diff + mn_now;
+ ev_rt_now = rtmn_diff + mn_now;
return 0;
}
else
{
now_floor = mn_now;
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
return 1;
}
}
for (i = 4; --i; ) /* loop a few times, before making important decisions */
{
- rtmn_diff = rt_now - mn_now;
+ rtmn_diff = ev_rt_now - mn_now;
if (fabs (odiff - rtmn_diff) < MIN_TIMEJUMP)
return; /* all is well */
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
mn_now = get_clock ();
now_floor = mn_now;
}
else
#endif
{
- rt_now = ev_time ();
+ ev_rt_now = ev_time ();
- if (expect_false (mn_now > rt_now || mn_now < rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP))
+ if (expect_false (mn_now > ev_rt_now || mn_now < ev_rt_now - MAX_BLOCKTIME - MIN_TIMEJUMP))
{
periodics_reschedule (EV_A);
/* adjust timers. this is easy, as the offset is the same for all */
for (i = 0; i < timercnt; ++i)
- ((WT)timers [i])->at += rt_now - mn_now;
+ ((WT)timers [i])->at += ev_rt_now - mn_now;
}
- mn_now = rt_now;
+ mn_now = ev_rt_now;
}
}
else
#endif
{
- rt_now = ev_time ();
- mn_now = rt_now;
+ ev_rt_now = ev_time ();
+ mn_now = ev_rt_now;
}
if (flags & EVLOOP_NONBLOCK || idlecnt)
if (periodiccnt)
{
- ev_tstamp to = ((WT)periodics [0])->at - rt_now + method_fudge;
+ ev_tstamp to = ((WT)periodics [0])->at - ev_rt_now + method_fudge;
if (block > to) block = to;
}
method_poll (EV_A_ block);
- /* update rt_now, do magic */
+ /* update ev_rt_now, do magic */
time_update (EV_A);
/* queue pending timers and reschedule them */
if (ev_is_active (w))
{
if (w->repeat)
- {
- ((WT)w)->at = mn_now + w->repeat;
- downheap ((WT *)timers, timercnt, ((W)w)->active - 1);
- }
+ adjustheap ((WT *)timers, timercnt, ((W)w)->active - 1, mn_now + w->repeat);
else
ev_timer_stop (EV_A_ w);
}
return;
if (w->reschedule_cb)
- ((WT)w)->at = w->reschedule_cb (w, rt_now);
+ ((WT)w)->at = w->reschedule_cb (w, ev_rt_now);
else if (w->interval)
{
assert (("ev_periodic_start called with negative interval value", w->interval >= 0.));
/* this formula differs from the one in periodic_reify because we do not always round up */
- ((WT)w)->at += ceil ((rt_now - ((WT)w)->at) / w->interval) * w->interval;
+ ((WT)w)->at += ceil ((ev_rt_now - ((WT)w)->at) / w->interval) * w->interval;
}
ev_start (EV_A_ (W)w, ++periodiccnt);
void
ev_periodic_again (EV_P_ struct ev_periodic *w)
{
+ /* TODO: use adjustheap and recalculation */
ev_periodic_stop (EV_A_ w);
ev_periodic_start (EV_A_ w);
}
once->cb = cb;
once->arg = arg;
- ev_watcher_init (&once->io, once_cb_io);
+ ev_init (&once->io, once_cb_io);
if (fd >= 0)
{
ev_io_set (&once->io, fd, events);
ev_io_start (EV_A_ &once->io);
}
- ev_watcher_init (&once->to, once_cb_to);
+ ev_init (&once->to, once_cb_to);
if (timeout >= 0.)
{
ev_timer_set (&once->to, timeout, 0.);
}
}
+#ifdef __cplusplus
+}
+#endif
+