From 77a6c3512a44dfe6540dde71946e6484fe4f173f Mon Sep 17 00:00:00 2001 From: lichao <lichao@aiotlink.com> Date: 星期一, 10 五月 2021 16:05:28 +0800 Subject: [PATCH] test lock code. --- src/socket.cpp | 181 +++++++++++++++++++++++++------------------- 1 files changed, 103 insertions(+), 78 deletions(-) diff --git a/src/socket.cpp b/src/socket.cpp index 5eb6756..6231579 100644 --- a/src/socket.cpp +++ b/src/socket.cpp @@ -24,104 +24,113 @@ using namespace bhome_msg; using namespace bhome_shm; -namespace +ShmSocket::ShmSocket(Shm &shm, const MQId id, const int len) : + run_(false), mq_(id, shm, len) { - -} // namespace - -ShmSocket::ShmSocket(Type type, bhome_shm::SharedMemory &shm) : - shm_(shm), type_(type), run_(false) -{ - switch (type) { - case eSockBus: mq_.reset(new Queue(kBHBusQueueId, shm_, 1000)); break; - case eSockRequest: mq_.reset(new Queue(shm_, 12)); break; - case eSockReply: mq_.reset(new Queue(shm_, 64)); break; - case eSockSubscribe: mq_.reset(new Queue(shm_, 64)); break; - case eSockPublish: break; // no recv mq needed - default: break; - } + Start(); } - -ShmSocket::ShmSocket(Type type) : - ShmSocket(type, BHomeShm()) {} +ShmSocket::ShmSocket(Shm &shm, const bool create_or_else_find, const MQId id, const int len) : + run_(false), mq_(id, create_or_else_find, shm, len) +{ + Start(); +} +ShmSocket::ShmSocket(bhome_shm::SharedMemory &shm, const int len) : + run_(false), mq_(shm, len) +{ + Start(); +} ShmSocket::~ShmSocket() { Stop(); } -bool ShmSocket::Publish(const std::string &topic, const void *data, const size_t size, const int timeout_ms) +bool ShmSocket::Start(const RawRecvCB &onData, const IdleCB &onIdle, int nworker) { - if (type_ != eSockPublish) { - return false; - } - assert(!mq_); - try { - MsgI imsg; - if (!imsg.MakeRC(shm_, MakePub(topic, data, size))) { - return false; - } - DEFER1(imsg.Release(shm_)); - return Queue::Send(shm_, kBHBusQueueId, imsg, timeout_ms); - - } catch (...) { - return false; - } -} - -bool ShmSocket::Subscribe(const std::vector<std::string> &topics, const int timeout_ms) -{ - if (type_ != eSockSubscribe) { - return false; - } - assert(mq_); - try { - return mq_->Send(kBHBusQueueId, MakeSub(mq_->Id(), topics), timeout_ms); - } catch (...) { - return false; - } -} - -bool ShmSocket::StartRaw(const RecvRawCB &onData, int nworker) -{ - auto CanRecv = [this]() { - switch (type_) { - case eSockRequest: - case eSockReply: - case eSockBus: - case eSockSubscribe: - return true; - default: - return false; - } - }; - if (!CanRecv()) { - return false; - } - std::lock_guard<std::mutex> lock(mutex_); - - StopNoLock(); - auto RecvProc = [this, onData]() { - while (run_) { - try { - MsgI imsg; - DEFER1(imsg.Release(shm_)); - if (mq_->Recv(imsg, 100)) { onData(imsg); } - } catch (...) { + auto ioProc = [this, onData, onIdle]() { + auto DoSend = [this]() { return send_buffer_.TrySend(mq()); }; + auto DoRecv = [=] { + // do not recv if no cb is set. + if (!onData) { + return false; } + auto onMsg = [&](MsgI &imsg) { + DEFER1(imsg.Release()); + onData(*this, imsg); + }; + MsgI imsg; + return mq().TryRecv(imsg) ? (onMsg(imsg), true) : false; + }; + + try { + bool more_to_send = DoSend(); + bool more_to_recv = DoRecv(); + if (onIdle) { onIdle(*this); } + if (!more_to_send && !more_to_recv) { + robust::QuickSleep(); + } + } catch (...) { } }; + + std::lock_guard<std::mutex> lock(mutex_); + StopNoLock(); run_.store(true); for (int i = 0; i < nworker; ++i) { - workers_.emplace_back(RecvProc); + workers_.emplace_back([this, ioProc]() { while (run_) { ioProc(); } }); } return true; } -bool ShmSocket::Start(const RecvCB &onData, int nworker) +bool ShmSocket::Start(int nworker, const RecvCB &onData, const IdleCB &onIdle) { - return StartRaw([this, onData](MsgI &imsg) { BHMsg m; if (imsg.Unpack(m)) { onData(m); } }, nworker); + auto ioProc = [this, onData, onIdle]() { + auto DoSend = [this]() { return send_buffer_.TrySend(mq()); }; + auto DoRecv = [=] { + auto onRecvWithPerMsgCB = [this, onData](ShmSocket &socket, MsgI &imsg, BHMsgHead &head) { + RecvCB cb; + if (per_msg_cbs_->Pick(head.msg_id(), cb)) { + cb(socket, imsg, head); + } else if (onData) { + onData(socket, imsg, head); + } + }; + + // do not recv if no cb is set. + if (!onData && per_msg_cbs_->empty()) { + return false; + } + auto onMsg = [&](MsgI &imsg) { + DEFER1(imsg.Release()); + BHMsgHead head; + if (imsg.ParseHead(head)) { + onRecvWithPerMsgCB(*this, imsg, head); + } + }; + MsgI imsg; + return mq().TryRecv(imsg) ? (onMsg(imsg), true) : false; + }; + + try { + bool more_to_send = DoSend(); + bool more_to_recv = DoRecv(); + if (onIdle) { onIdle(*this); } + if (!more_to_send && !more_to_recv) { + robust::QuickSleep(); + } + } catch (...) { + } + }; + + std::lock_guard<std::mutex> lock(mutex_); + StopNoLock(); + + run_.store(true); + for (int i = 0; i < nworker; ++i) { + workers_.emplace_back([this, ioProc]() { while (run_) { ioProc(); } }); + } + return true; } bool ShmSocket::Stop() @@ -138,7 +147,23 @@ w.join(); } } + workers_.clear(); return true; } return false; } + +//maybe reimplment, using async cbs? +bool ShmSocket::SyncRecv(bhome_msg::MsgI &msg, bhome_msg::BHMsgHead &head, const int timeout_ms) +{ + // std::lock_guard<std::mutex> lock(mutex_); // seems no need to lock mutex_. + bool got = (timeout_ms == 0) ? mq().TryRecv(msg) : mq().Recv(msg, timeout_ms); + if (got) { + if (msg.ParseHead(head)) { + return true; + } else { + msg.Release(); + } + } + return false; +} -- Gitblit v1.8.0