From 02ba913dc7bb5d711471b27f2ea23a897d0f2e28 Mon Sep 17 00:00:00 2001 From: lichao <lichao@aiotlink.com> Date: 星期五, 23 四月 2021 15:34:26 +0800 Subject: [PATCH] bind msgi to shm, change offset_ptr to abs offset. --- utest/api_test.cpp | 393 ++++++++++++++++++++++++++++++++++++++++++++++++++++++-- 1 files changed, 379 insertions(+), 14 deletions(-) diff --git a/utest/api_test.cpp b/utest/api_test.cpp index 113bb99..200ae99 100644 --- a/utest/api_test.cpp +++ b/utest/api_test.cpp @@ -17,11 +17,253 @@ */ #include "bh_api.h" #include "util.h" +#include <atomic> -class DemoClient +using namespace bhome_msg; + +namespace { -public: +typedef std::atomic<uint64_t> Number; + +void Assign(Number &a, const Number &b) { a.store(b.load()); } +struct MsgStatus { + + Number nrequest_; + Number nfailed_; + Number nreply_; + Number nserved_; + MsgStatus() : + nrequest_(0), nreply_(0), nserved_(0) {} + MsgStatus &operator=(const MsgStatus &a) + { + Assign(nrequest_, a.nrequest_); + Assign(nserved_, a.nserved_); + Assign(nreply_, a.nreply_); + Assign(nfailed_, a.nfailed_); + return *this; + } }; + +MsgStatus &Status() +{ + static MsgStatus st; + return st; +} +} // namespace + +void SubRecvProc(const void *proc_id, + const int proc_id_len, + const void *data, + const int data_len) +{ + std::string proc((const char *) proc_id, proc_id_len); + MsgPublish pub; + pub.ParseFromArray(data, data_len); + // printf("Sub data, %s : %s\n", pub.topic().c_str(), pub.data().c_str()); +} + +void ServerProc(const void *proc_id, + const int proc_id_len, + const void *data, + const int data_len, + void *src) +{ + // printf("ServerProc: "); + // DEFER1(printf("\n");); + MsgRequestTopic request; + if (request.ParseFromArray(data, data_len)) { + MsgRequestTopicReply reply; + reply.set_data(" reply: " + request.data()); + std::string s(reply.SerializeAsString()); + // printf("%s", reply.data().c_str()); + BHSendReply(src, s.data(), s.size()); + ++Status().nserved_; + } +} + +void ClientProc(const void *proc_id, + const int proc_id_len, + const void *msg_id, + const int msg_id_len, + const void *data, + const int data_len) +{ + std::string proc((const char *) proc_id, proc_id_len); + MsgRequestTopicReply reply; + if (reply.ParseFromArray(data, data_len)) { + ++Status().nreply_; + } + // printf("client Recv reply : %s\n", reply.data().c_str()); +} + +class TLMutex +{ + // typedef boost::interprocess::interprocess_mutex MutexT; + typedef CasMutex MutexT; + // typedef std::mutex MutexT; + typedef std::chrono::steady_clock Clock; + typedef Clock::duration Duration; + static Duration Now() { return Clock::now().time_since_epoch(); } + + const Duration limit_; + std::atomic<Duration> last_lock_time_; + MutexT mutex_; + +public: + struct Status { + int64_t nlock_ = 0; + int64_t nupdate_time_fail = 0; + int64_t nfail = 0; + int64_t nexcept = 0; + }; + Status st_; + + explicit TLMutex(Duration limit) : + limit_(limit) {} + TLMutex() : + TLMutex(std::chrono::seconds(1)) {} + ~TLMutex() { static_assert(std::is_pod<Duration>::value); } + bool try_lock() + { + if (mutex_.try_lock()) { + auto old_time = last_lock_time_.load(); + if (Now() - old_time > limit_) { + return last_lock_time_.compare_exchange_strong(old_time, Now()); + } else { + last_lock_time_.store(Now()); + return true; + } + } else { + auto old_time = last_lock_time_.load(); + if (Now() - old_time > limit_) { + return last_lock_time_.compare_exchange_strong(old_time, Now()); + } else { + return false; + } + } + } + void lock() + { + int n = 0; + while (!try_lock()) { + n++; + std::this_thread::yield(); + } + st_.nlock_ += n; + } + void unlock() { mutex_.unlock(); } +}; + +namespace +{ +typedef int64_t Offset; +Offset Addr(void *ptr) { return reinterpret_cast<Offset>(ptr); } +void *Ptr(const Offset offset) { return reinterpret_cast<void *>(offset); } +} // namespace + +BOOST_AUTO_TEST_CASE(MutexTest) +{ + SharedMemory &shm = TestShm(); + MsgI::BindShm(shm); + + void *base_ptr = shm.get_address(); + auto PrintPtr = [&](void *p) { + printf("addr: %ld, ptr: %p, offset: %ld\n", Addr(p), p, Addr(p) - Addr(base_ptr)); + }; + + printf("base"); + PrintPtr(base_ptr); + + MsgI msg; + msg.Make("string data"); + for (int i = 0; i < 10; ++i) { + int n = msg.AddRef(); + printf("add %d ref: %d\n", i, n); + } + for (int i = 0; i < 10; ++i) { + int n = msg.Release(); + printf("release %d, ref : %d\n", i, n); + } + std::this_thread::sleep_for(1s); + msg.Release(); + + const std::string mtx_name("test_mutex"); + const std::string int_name("test_int"); + auto mtx = shm.find_or_construct<Mutex>(mtx_name.c_str())(); + auto pi = shm.find_or_construct<int>(int_name.c_str())(100); + + printf("mutetx "); + PrintPtr(mtx); + printf("int "); + PrintPtr(pi); + + typedef std::chrono::steady_clock Clock; + auto Now = []() { return Clock::now().time_since_epoch(); }; + if (pi) { + auto old = *pi; + printf("int : %d, add1: %d\n", old, ++*pi); + } + + { + boost::timer::auto_cpu_timer timer; + printf("test time: "); + TLMutex mutex; + // CasMutex mutex; + auto Lock = [&]() { + for (int i = 0; i < 1000 * 100; ++i) { + mutex.lock(); + mutex.unlock(); + } + }; + std::thread t1(Lock), t2(Lock); + t1.join(); + t2.join(); + printf("mutex nlock: %ld, update time error: %ld, normal fail: %ld, error wait: %ld\n", + mutex.st_.nlock_, + mutex.st_.nupdate_time_fail, + mutex.st_.nfail, + mutex.st_.nexcept); + } + + auto MSFromNow = [](const int ms) { + using namespace boost::posix_time; + ptime cur = boost::posix_time::microsec_clock::universal_time(); + return cur + millisec(ms); + }; + + auto TryLock = [&]() { + if (mtx->try_lock()) { + printf("try_lock ok\n"); + return true; + } else { + printf("try_lock failed\n"); + return false; + } + }; + auto Unlock = [&]() { + mtx->unlock(); + printf("unlocked\n"); + }; + + if (mtx) { + printf("mtx exists\n"); + if (TryLock()) { + auto op = [&]() { + if (TryLock()) { + Unlock(); + } + }; + op(); + std::thread t(op); + t.join(); + // Unlock(); + } else { + // mtx->unlock(); + } + } else { + printf("mtx not exists\n"); + } +} BOOST_AUTO_TEST_CASE(ApiTest) { @@ -36,19 +278,142 @@ nsec, nhour, nday, years); std::chrono::steady_clock::duration a(123456); printf("nowsec: %ld\n", NowSec()); - // for (int i = 0; i < 5; ++i) { - // std::this_thread::sleep_for(1s); - // printf("nowsec: %ld\n", NowSec()); - // } printf("maxsec: %ld\n", CountSeconds(max_time)); - ProcInfo proc; - proc.set_proc_id("demo_client"); - proc.set_public_info("public info of demo_client. etc..."); - std::string proc_buf(proc.SerializeAsString()); - void *reply = 0; - int reply_len = 0; - bool r = BHRegister(proc_buf.data(), proc_buf.size(), &reply, &reply_len, 1000); - printf("register %s\n", r ? "ok" : "failed"); + bool reg = false; + for (int i = 0; i < 3 && !reg; ++i) { + ProcInfo proc; + proc.set_proc_id("demo_client"); + proc.set_public_info("public info of demo_client. etc..."); + std::string proc_buf(proc.SerializeAsString()); + void *reply = 0; + int reply_len = 0; + reg = BHRegister(proc_buf.data(), proc_buf.size(), &reply, &reply_len, 2000); + printf("register %s\n", reg ? "ok" : "failed"); + + BHFree(reply, reply_len); + Sleep(1s); + } + if (!reg) { + return; + } + + const std::string topic_ = "topic_"; + + { + MsgTopicList topics; + for (int i = 0; i < 10; ++i) { + topics.add_topic_list(topic_ + std::to_string(i)); + } + std::string s = topics.SerializeAsString(); + void *reply = 0; + int reply_len = 0; + bool r = BHRegisterTopics(s.data(), s.size(), &reply, &reply_len, 1000); + BHFree(reply, reply_len); + // printf("register topic : %s\n", r ? "ok" : "failed"); + Sleep(1s); + } + + { + MsgTopicList topics; + for (int i = 0; i < 10; ++i) { + topics.add_topic_list(topic_ + std::to_string(i * 2)); + } + std::string s = topics.SerializeAsString(); + void *reply = 0; + int reply_len = 0; + bool r = BHSubscribeTopics(s.data(), s.size(), &reply, &reply_len, 1000); + BHFree(reply, reply_len); + printf("subscribe topic : %s\n", r ? "ok" : "failed"); + } + + BHStartWorker(&ServerProc, &SubRecvProc, &ClientProc); + + { + for (int i = 0; i < 1; ++i) { + MsgPublish pub; + pub.set_topic(topic_ + std::to_string(i)); + pub.set_data("pub_data_" + std::string(1024 * 1, 'a')); + std::string s(pub.SerializeAsString()); + BHPublish(s.data(), s.size(), 0); + // Sleep(1s); + } + } + + auto asyncRequest = [&](uint64_t nreq) { + for (uint64_t i = 0; i < nreq; ++i) { + MsgRequestTopic req; + req.set_topic(topic_ + std::to_string(0)); + req.set_data("request_data_" + std::to_string(i)); + std::string s(req.SerializeAsString()); + void *msg_id = 0; + int len = 0; + // Sleep(10ms, false); + std::string dest(BHAddress().SerializeAsString()); + bool r = BHAsyncRequest(dest.data(), dest.size(), s.data(), s.size(), 0, 0); + DEFER1(BHFree(msg_id, len);); + if (r) { + ++Status().nrequest_; + } else { + ++Status().nfailed_; + static std::atomic<int64_t> last(0); + auto now = NowSec(); + if (last.exchange(now) < now) { + int ec = 0; + std::string msg; + GetLastError(ec, msg); + printf("request topic error --------- : %s\n", msg.c_str()); + } + } + } + }; + auto showStatus = [](std::atomic<bool> *run) { + MsgStatus last; + while (*run) { + auto &st = Status(); + Sleep(1s, false); + printf("nreq: %8ld, spd %8ld | failed: %8ld | nsrv: %8ld, spd %8ld | nreply: %8ld, spd %8ld\n", + st.nrequest_.load(), st.nrequest_ - last.nrequest_, + st.nfailed_.load(), + st.nserved_.load(), st.nserved_ - last.nserved_, + st.nreply_.load(), st.nreply_ - last.nreply_); + last = st; + } + }; + auto hb = [](std::atomic<bool> *run) { + while (*run) { + Sleep(1s, false); + bool r = BHHeartbeatEasy(1000); + printf("heartbeat: %s\n", r ? "ok" : "failed"); + } + }; + std::atomic<bool> run(true); + ThreadManager threads; + boost::timer::auto_cpu_timer timer; + threads.Launch(hb, &run); + threads.Launch(showStatus, &run); + int ncli = 10; + const uint64_t nreq = 1000 * 100; + for (int i = 0; i < ncli; ++i) { + threads.Launch(asyncRequest, nreq); + } + + int same = 0; + int64_t last = 0; + while (last < nreq * ncli && same < 2) { + Sleep(1s, false); + auto cur = Status().nreply_.load(); + if (last == cur) { + ++same; + } else { + last = cur; + same = 0; + } + } + + run = false; + threads.WaitAll(); + auto &st = Status(); + printf("nreq: %8ld, nsrv: %8ld, nreply: %8ld\n", st.nrequest_.load(), st.nserved_.load(), st.nreply_.load()); } \ No newline at end of file -- Gitblit v1.8.0