255 lines
7.3 KiB
D
255 lines
7.3 KiB
D
module eventcore.drivers.posix.events;
|
|
@safe:
|
|
|
|
import eventcore.driver;
|
|
import eventcore.drivers.posix.driver;
|
|
import eventcore.internal.consumablequeue : ConsumableQueue;
|
|
|
|
version (linux) {
|
|
nothrow @nogc extern (C) int eventfd(uint initval, int flags);
|
|
enum EFD_NONBLOCK = 0x800;
|
|
enum EFD_CLOEXEC = 0x80000;
|
|
}
|
|
version (Posix) {
|
|
import core.sys.posix.unistd : close, read, write;
|
|
} else {
|
|
import core.sys.windows.winsock2 : closesocket, AF_INET, SOCKET, SOCK_DGRAM,
|
|
bind, connect, getsockname, send, socket;
|
|
}
|
|
|
|
|
|
final class PosixEventDriverEvents(Loop : PosixEventLoop, Sockets : EventDriverSockets) : EventDriverEvents {
|
|
@safe: /*@nogc:*/ nothrow:
|
|
private {
|
|
Loop m_loop;
|
|
Sockets m_sockets;
|
|
ubyte[ulong.sizeof] m_buf;
|
|
version (linux) {}
|
|
else {
|
|
// TODO: avoid the overhead of a mutex backed map here
|
|
import core.sync.mutex : Mutex;
|
|
Mutex m_eventsMutex;
|
|
EventID[DatagramSocketFD] m_events;
|
|
}
|
|
}
|
|
|
|
this(Loop loop, Sockets sockets)
|
|
{
|
|
m_loop = loop;
|
|
m_sockets = sockets;
|
|
version (linux) {}
|
|
else m_eventsMutex = new Mutex;
|
|
}
|
|
|
|
package @property Loop loop() { return m_loop; }
|
|
|
|
final override EventID create()
|
|
{
|
|
return createInternal(false);
|
|
}
|
|
|
|
package(eventcore) EventID createInternal(bool is_internal = true)
|
|
{
|
|
version (linux) {
|
|
auto eid = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
|
|
if (eid == -1) return EventID.invalid;
|
|
auto id = cast(EventID)eid;
|
|
m_loop.initFD(id, FDFlags.internal);
|
|
m_loop.m_fds[id].specific = EventSlot(new ConsumableQueue!EventCallback, false, is_internal); // FIXME: avoid dynamic memory allocation
|
|
m_loop.registerFD(id, EventMask.read);
|
|
m_loop.setNotifyCallback!(EventType.read)(id, &onEvent);
|
|
releaseRef(id); // setNotifyCallback increments the reference count, but we need a value of 1 upon return
|
|
assert(getRC(id) == 1);
|
|
return id;
|
|
} else {
|
|
sock_t[2] fd;
|
|
version (Posix) {
|
|
import core.sys.posix.fcntl : fcntl, F_SETFL;
|
|
import eventcore.drivers.posix.sockets : O_CLOEXEC;
|
|
|
|
// create a pair of sockets to communicate between threads
|
|
import core.sys.posix.sys.socket : SOCK_DGRAM, AF_UNIX, socketpair;
|
|
if (() @trusted { return socketpair(AF_UNIX, SOCK_DGRAM, 0, fd); } () != 0)
|
|
return EventID.invalid;
|
|
|
|
assert(fd[0] != fd[1]);
|
|
|
|
// use the first socket as the async receiver
|
|
auto s = m_sockets.adoptDatagramSocketInternal(fd[0], true, true);
|
|
|
|
() @trusted { fcntl(fd[1], F_SETFL, O_CLOEXEC); } ();
|
|
} else {
|
|
// fake missing socketpair support on Windows
|
|
import std.socket : InternetAddress;
|
|
auto addr = new InternetAddress(0x7F000001, 0);
|
|
auto s = m_sockets.createDatagramSocketInternal(addr, null, true);
|
|
if (s == DatagramSocketFD.invalid) return EventID.invalid;
|
|
fd[0] = cast(sock_t)s;
|
|
if (!() @trusted {
|
|
fd[1] = socket(AF_INET, SOCK_DGRAM, 0);
|
|
int nl = addr.nameLen;
|
|
import eventcore.internal.utils : print;
|
|
if (bind(fd[1], addr.name, addr.nameLen) != 0)
|
|
return false;
|
|
assert(nl == addr.nameLen);
|
|
if (getsockname(fd[0], addr.name, &nl) != 0)
|
|
return false;
|
|
if (connect(fd[1], addr.name, addr.nameLen) != 0)
|
|
return false;
|
|
return true;
|
|
} ())
|
|
{
|
|
m_sockets.releaseRef(s);
|
|
return EventID.invalid;
|
|
}
|
|
}
|
|
|
|
m_sockets.receive(s, m_buf, IOMode.once, &onSocketData);
|
|
|
|
// use the second socket as the event ID and as the sending end for
|
|
// other threads
|
|
auto id = cast(EventID)fd[1];
|
|
try {
|
|
synchronized (m_eventsMutex)
|
|
m_events[s] = id;
|
|
} catch (Exception e) assert(false, e.msg);
|
|
m_loop.initFD(id, FDFlags.internal);
|
|
m_loop.m_fds[id].specific = EventSlot(new ConsumableQueue!EventCallback, false, is_internal, s); // FIXME: avoid dynamic memory allocation
|
|
assert(getRC(id) == 1);
|
|
return id;
|
|
}
|
|
}
|
|
|
|
final override void trigger(EventID event, bool notify_all)
|
|
{
|
|
auto slot = getSlot(event);
|
|
if (notify_all) {
|
|
//log("emitting only for this thread (%s waiters)", m_fds[event].waiters.length);
|
|
foreach (w; slot.waiters.consume) {
|
|
//log("emitting waiter %s %s", cast(void*)w.funcptr, w.ptr);
|
|
if (!isInternal(event)) m_loop.m_waiterCount--;
|
|
w(event);
|
|
}
|
|
} else {
|
|
if (!slot.waiters.empty) {
|
|
if (!isInternal(event)) m_loop.m_waiterCount--;
|
|
slot.waiters.consumeOne()(event);
|
|
}
|
|
}
|
|
}
|
|
|
|
final override void trigger(EventID event, bool notify_all)
|
|
shared @trusted {
|
|
import core.atomic : atomicStore;
|
|
auto thisus = cast(PosixEventDriverEvents)this;
|
|
assert(event < thisus.m_loop.m_fds.length, "Invalid event ID passed to shared triggerEvent.");
|
|
long one = 1;
|
|
//log("emitting for all threads");
|
|
if (notify_all) atomicStore(thisus.getSlot(event).triggerAll, true);
|
|
version (Posix) .write(cast(int)event, &one, one.sizeof);
|
|
else assert(send(cast(int)event, cast(const(ubyte*))&one, one.sizeof, 0) == one.sizeof);
|
|
}
|
|
|
|
final override void wait(EventID event, EventCallback on_event)
|
|
{
|
|
if (!isInternal(event)) m_loop.m_waiterCount++;
|
|
getSlot(event).waiters.put(on_event);
|
|
}
|
|
|
|
final override void cancelWait(EventID event, EventCallback on_event)
|
|
{
|
|
import std.algorithm.searching : countUntil;
|
|
import std.algorithm.mutation : remove;
|
|
|
|
if (!isInternal(event)) m_loop.m_waiterCount--;
|
|
getSlot(event).waiters.removePending(on_event);
|
|
}
|
|
|
|
private void onEvent(FD fd)
|
|
@trusted {
|
|
EventID event = cast(EventID)fd;
|
|
version (linux) {
|
|
ulong cnt;
|
|
() @trusted { .read(cast(int)event, &cnt, cnt.sizeof); } ();
|
|
}
|
|
import core.atomic : cas;
|
|
auto all = cas(&getSlot(event).triggerAll, true, false);
|
|
trigger(event, all);
|
|
}
|
|
|
|
version (linux) {}
|
|
else {
|
|
private void onSocketData(DatagramSocketFD s, IOStatus, size_t, scope RefAddress)
|
|
{
|
|
m_sockets.receive(s, m_buf, IOMode.once, &onSocketData);
|
|
EventID evt;
|
|
try {
|
|
synchronized (m_eventsMutex)
|
|
evt = m_events[s];
|
|
onEvent(evt);
|
|
} catch (Exception e) assert(false, e.msg);
|
|
}
|
|
}
|
|
|
|
final override void addRef(EventID descriptor)
|
|
{
|
|
assert(getRC(descriptor) > 0, "Adding reference to unreferenced event FD.");
|
|
getRC(descriptor)++;
|
|
}
|
|
|
|
final override bool releaseRef(EventID descriptor)
|
|
{
|
|
assert(getRC(descriptor) > 0, "Releasing reference to unreferenced event FD.");
|
|
if (--getRC(descriptor) == 0) {
|
|
if (!isInternal(descriptor))
|
|
m_loop.m_waiterCount -= getSlot(descriptor).waiters.length;
|
|
() @trusted nothrow {
|
|
try .destroy(getSlot(descriptor).waiters);
|
|
catch (Exception e) assert(false, e.msg);
|
|
} ();
|
|
version (linux) {
|
|
m_loop.unregisterFD(descriptor, EventMask.read);
|
|
} else {
|
|
auto rs = getSlot(descriptor).recvSocket;
|
|
m_sockets.cancelReceive(rs);
|
|
m_sockets.releaseRef(rs);
|
|
try {
|
|
synchronized (m_eventsMutex)
|
|
m_events.remove(rs);
|
|
} catch (Exception e) assert(false, e.msg);
|
|
}
|
|
m_loop.clearFD(descriptor);
|
|
version (Posix) close(cast(int)descriptor);
|
|
else () @trusted { closesocket(cast(SOCKET)descriptor); } ();
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
private EventSlot* getSlot(EventID id)
|
|
{
|
|
assert(id < m_loop.m_fds.length, "Invalid event ID.");
|
|
return () @trusted { return &m_loop.m_fds[id].event(); } ();
|
|
}
|
|
|
|
private ref uint getRC(EventID id)
|
|
{
|
|
return m_loop.m_fds[id].common.refCount;
|
|
}
|
|
|
|
private bool isInternal(EventID id)
|
|
{
|
|
return getSlot(id).isInternal;
|
|
}
|
|
}
|
|
|
|
package struct EventSlot {
|
|
alias Handle = EventID;
|
|
ConsumableQueue!EventCallback waiters;
|
|
shared bool triggerAll;
|
|
bool isInternal;
|
|
version (linux) {}
|
|
else {
|
|
DatagramSocketFD recvSocket;
|
|
}
|
|
}
|