From a6f67b4249525089fb97eb9418c7014f66c2a000 Mon Sep 17 00:00:00 2001 From: lichao <lichao@aiotlink.com> Date: 星期三, 28 四月 2021 19:29:17 +0800 Subject: [PATCH] use new robust mutex, circurar; rm timeout mutex. --- utest/api_test.cpp | 192 +++++++++++++++++++++-------------------------- 1 files changed, 85 insertions(+), 107 deletions(-) diff --git a/utest/api_test.cpp b/utest/api_test.cpp index 200ae99..6577b51 100644 --- a/utest/api_test.cpp +++ b/utest/api_test.cpp @@ -16,8 +16,10 @@ * ===================================================================================== */ #include "bh_api.h" +#include "robust.h" #include "util.h" #include <atomic> +#include <boost/lockfree/queue.hpp> using namespace bhome_msg; @@ -49,7 +51,6 @@ static MsgStatus st; return st; } -} // namespace void SubRecvProc(const void *proc_id, const int proc_id_len, @@ -59,7 +60,7 @@ std::string proc((const char *) proc_id, proc_id_len); MsgPublish pub; pub.ParseFromArray(data, data_len); - // printf("Sub data, %s : %s\n", pub.topic().c_str(), pub.data().c_str()); + printf("Sub data, %s : %s\n", pub.topic().c_str(), pub.data().c_str()); } void ServerProc(const void *proc_id, @@ -96,107 +97,33 @@ // printf("client Recv reply : %s\n", reply.data().c_str()); } -class TLMutex -{ - // typedef boost::interprocess::interprocess_mutex MutexT; - typedef CasMutex MutexT; - // typedef std::mutex MutexT; - typedef std::chrono::steady_clock Clock; - typedef Clock::duration Duration; - static Duration Now() { return Clock::now().time_since_epoch(); } - - const Duration limit_; - std::atomic<Duration> last_lock_time_; - MutexT mutex_; - -public: - struct Status { - int64_t nlock_ = 0; - int64_t nupdate_time_fail = 0; - int64_t nfail = 0; - int64_t nexcept = 0; - }; - Status st_; - - explicit TLMutex(Duration limit) : - limit_(limit) {} - TLMutex() : - TLMutex(std::chrono::seconds(1)) {} - ~TLMutex() { static_assert(std::is_pod<Duration>::value); } - bool try_lock() - { - if (mutex_.try_lock()) { - auto old_time = last_lock_time_.load(); - if (Now() - old_time > limit_) { - return last_lock_time_.compare_exchange_strong(old_time, Now()); - } else { - last_lock_time_.store(Now()); - return true; - } - } else { - auto old_time = last_lock_time_.load(); - if (Now() - old_time > limit_) { - return last_lock_time_.compare_exchange_strong(old_time, Now()); - } else { - return false; - } - } - } - void lock() - { - int n = 0; - while (!try_lock()) { - n++; - std::this_thread::yield(); - } - st_.nlock_ += n; - } - void unlock() { mutex_.unlock(); } -}; - -namespace -{ -typedef int64_t Offset; -Offset Addr(void *ptr) { return reinterpret_cast<Offset>(ptr); } -void *Ptr(const Offset offset) { return reinterpret_cast<void *>(offset); } } // namespace +#include <chrono> +using namespace std::chrono; +// using namespace std::chrono_literals; BOOST_AUTO_TEST_CASE(MutexTest) { + // typedef robust::CasMutex<true> RobustMutex; + typedef MutexWithPidCheck RobustMutex; + + for (int i = 0; i < 20; ++i) { + int size = i; + int left = size & 7; + int rsize = size + ((8 - left) & 7); + printf("size: %3d, rsize: %3d\n", size, rsize); + } SharedMemory &shm = TestShm(); - MsgI::BindShm(shm); - - void *base_ptr = shm.get_address(); - auto PrintPtr = [&](void *p) { - printf("addr: %ld, ptr: %p, offset: %ld\n", Addr(p), p, Addr(p) - Addr(base_ptr)); - }; - - printf("base"); - PrintPtr(base_ptr); - - MsgI msg; - msg.Make("string data"); - for (int i = 0; i < 10; ++i) { - int n = msg.AddRef(); - printf("add %d ref: %d\n", i, n); - } - for (int i = 0; i < 10; ++i) { - int n = msg.Release(); - printf("release %d, ref : %d\n", i, n); - } - std::this_thread::sleep_for(1s); - msg.Release(); + // shm.Remove(); + // return; + GlobalInit(shm); const std::string mtx_name("test_mutex"); const std::string int_name("test_int"); - auto mtx = shm.find_or_construct<Mutex>(mtx_name.c_str())(); - auto pi = shm.find_or_construct<int>(int_name.c_str())(100); + auto mtx = shm.FindOrCreate<RobustMutex>(mtx_name); + auto pi = shm.FindOrCreate<int>(int_name, 100); - printf("mutetx "); - PrintPtr(mtx); - printf("int "); - PrintPtr(pi); - + std::mutex m; typedef std::chrono::steady_clock Clock; auto Now = []() { return Clock::now().time_since_epoch(); }; if (pi) { @@ -206,11 +133,11 @@ { boost::timer::auto_cpu_timer timer; - printf("test time: "); - TLMutex mutex; - // CasMutex mutex; + const int ntimes = 1000 * 1000; + printf("test lock/unlock %d times: ", ntimes); + RobustMutex mutex; auto Lock = [&]() { - for (int i = 0; i < 1000 * 100; ++i) { + for (int i = 0; i < ntimes; ++i) { mutex.lock(); mutex.unlock(); } @@ -218,11 +145,6 @@ std::thread t1(Lock), t2(Lock); t1.join(); t2.join(); - printf("mutex nlock: %ld, update time error: %ld, normal fail: %ld, error wait: %ld\n", - mutex.st_.nlock_, - mutex.st_.nupdate_time_fail, - mutex.st_.nfail, - mutex.st_.nexcept); } auto MSFromNow = [](const int ms) { @@ -301,7 +223,7 @@ const std::string topic_ = "topic_"; - { + { // Server Register Topics MsgTopicList topics; for (int i = 0; i < 10; ++i) { topics.add_topic_list(topic_ + std::to_string(i)); @@ -315,7 +237,7 @@ Sleep(1s); } - { + { // Subscribe MsgTopicList topics; for (int i = 0; i < 10; ++i) { topics.add_topic_list(topic_ + std::to_string(i * 2)); @@ -328,8 +250,56 @@ printf("subscribe topic : %s\n", r ? "ok" : "failed"); } - BHStartWorker(&ServerProc, &SubRecvProc, &ClientProc); + auto ServerLoop = [&](std::atomic<bool> *run) { + while (*run) { + void *proc_id = 0; + int proc_id_len = 0; + DEFER1(BHFree(proc_id, proc_id_len);); + void *input = 0; + int input_len = 0; + DEFER1(BHFree(input, input_len)); + void *src = 0; + if (BHReadRequest(&proc_id, &proc_id_len, &input, &input_len, &src, 10) && src) { + MsgRequestTopic request; + if (request.ParseFromArray(input, input_len)) { + MsgRequestTopicReply reply; + reply.set_data(" reply: " + request.data()); + std::string s(reply.SerializeAsString()); + // printf("%s", reply.data().c_str()); + BHSendReply(src, s.data(), s.size()); + ++Status().nserved_; + } + src = 0; + } + } + }; + + auto SyncRequest = [&](int idx) { // SyncRequest + MsgRequestTopic req; + req.set_topic(topic_ + std::to_string(idx)); + req.set_data("request_data_" + std::to_string(idx)); + std::string s(req.SerializeAsString()); + // Sleep(10ms, false); + std::string dest(BHAddress().SerializeAsString()); + void *proc_id = 0; + int proc_id_len = 0; + DEFER1(BHFree(proc_id, proc_id_len);); + void *reply = 0; + int reply_len = 0; + DEFER1(BHFree(reply, reply_len)); + bool r = BHRequest(dest.data(), dest.size(), s.data(), s.size(), &proc_id, &proc_id_len, &reply, &reply_len, 100); + if (!r) { + int ec = 0; + std::string msg; + GetLastError(ec, msg); + printf("request error: %s\n", msg.c_str()); + } else { + MsgRequestTopicReply ret; + ret.ParseFromArray(reply, reply_len); + printf("request result: %s\n", ret.data().c_str()); + } + }; { for (int i = 0; i < 1; ++i) { MsgPublish pub; @@ -349,10 +319,10 @@ std::string s(req.SerializeAsString()); void *msg_id = 0; int len = 0; + DEFER1(BHFree(msg_id, len);); // Sleep(10ms, false); std::string dest(BHAddress().SerializeAsString()); bool r = BHAsyncRequest(dest.data(), dest.size(), s.data(), s.size(), 0, 0); - DEFER1(BHFree(msg_id, len);); if (r) { ++Status().nrequest_; } else { @@ -388,7 +358,10 @@ printf("heartbeat: %s\n", r ? "ok" : "failed"); } }; + std::atomic<bool> run(true); + + BHStartWorker(&ServerProc, &SubRecvProc, &ClientProc); ThreadManager threads; boost::timer::auto_cpu_timer timer; threads.Launch(hb, &run); @@ -398,6 +371,9 @@ for (int i = 0; i < ncli; ++i) { threads.Launch(asyncRequest, nreq); } + // for (int i = 0; i < 100; ++i) { + // SyncRequest(0); + // } int same = 0; int64_t last = 0; @@ -416,4 +392,6 @@ threads.WaitAll(); auto &st = Status(); printf("nreq: %8ld, nsrv: %8ld, nreply: %8ld\n", st.nrequest_.load(), st.nserved_.load(), st.nreply_.load()); + BHCleanup(); + printf("after cleanup\n"); } \ No newline at end of file -- Gitblit v1.8.0