From 3931f83205f153f2bc7fc36d1a894cdc3f14b4db Mon Sep 17 00:00:00 2001 From: lichao <lichao@aiotlink.com> Date: 星期三, 21 四月 2021 16:52:51 +0800 Subject: [PATCH] change node socket to vector; try lock free queue. --- utest/utest.cpp | 95 ++++++++++++++++++++++++++++++++++------------- 1 files changed, 69 insertions(+), 26 deletions(-) diff --git a/utest/utest.cpp b/utest/utest.cpp index e0a9023..b2de97f 100644 --- a/utest/utest.cpp +++ b/utest/utest.cpp @@ -1,6 +1,5 @@ #include "center.h" #include "defs.h" -#include "failed_msg.h" #include "util.h" #include <atomic> #include <boost/uuid/uuid_generators.hpp> @@ -21,8 +20,6 @@ struct IsSameType<A, A> { static const bool value = true; }; - -typedef FailedMsgQ ServerFailedQ; BOOST_AUTO_TEST_CASE(Temp) { @@ -102,10 +99,10 @@ BHCenter center(shm); center.Start(); - std::this_thread::sleep_for(100ms); + Sleep(100ms); std::atomic<uint64_t> total_count(0); - std::atomic<ptime> last_time(Now() - seconds(1)); + std::atomic<int64_t> last_time(NowSec() - 1); std::atomic<uint64_t> last_count(0); const uint64_t nmsg = 100 * 2; @@ -116,7 +113,8 @@ for (auto &t : topics) { tlist.add_topic_list(t); } - bool r = client.Subscribe(tlist, timeout); + MsgCommonReply reply_body; + bool r = client.Subscribe(tlist, reply_body, timeout); if (!r) { printf("client subscribe failed.\n"); } @@ -124,10 +122,10 @@ std::condition_variable cv; std::atomic<uint64_t> n(0); - auto OnTopicData = [&](const std::string &proc_id, const std::string &topic, const std::string &data) { + auto OnTopicData = [&](const std::string &proc_id, const MsgPublish &pub) { ++total_count; - auto cur = Now(); + auto cur = NowSec(); if (last_time.exchange(cur) < cur) { std::cout << "time: " << cur; printf("sub recv, total msg:%10ld, speed:[%8ld/s], used mem:%8ld \n", @@ -149,8 +147,10 @@ for (unsigned i = 0; i < nmsg; ++i) { std::string data = topic + std::to_string(i) + std::string(1000, '-'); - - bool r = provider.Publish(topic, data.data(), data.size(), timeout); + MsgPublish pub; + pub.set_topic(topic); + pub.set_data(data); + bool r = provider.Publish(pub, 0); if (!r) { static std::atomic<int> an(0); int n = ++an; @@ -170,14 +170,14 @@ part.push_back(topics[i]); threads.Launch(Sub, i, topics); } - std::this_thread::sleep_for(100ms); + Sleep(100ms); for (auto &topic : topics) { threads.Launch(Pub, topic); } threads.Launch(Pub, "some_else"); threads.WaitAll(); - std::cout << "end : " << Now(); + printf("sub recv, total msg:%10ld, speed:[%8ld/s], used mem:%8ld \n", total_count.load(), total_count - last_count.exchange(total_count), init_avail - Avail()); } @@ -198,7 +198,7 @@ { const std::string shm_name("ShmReqRep"); ShmRemover auto_remove(shm_name); - SharedMemory shm(shm_name, 1024 * 1024 * 50); + SharedMemory shm(shm_name, 1024 * 1024 * 512); auto Avail = [&]() { return shm.get_free_memory(); }; auto init_avail = Avail(); @@ -218,29 +218,37 @@ std::atomic<int> count(0); std::string reply; - auto onRecv = [&](const std::string &rep) { - reply = rep; + auto onRecv = [&](const BHMsgHead &head, const MsgRequestTopicReply &msg) { + reply = msg.data(); if (++count >= nreq) { printf("count: %d\n", count.load()); } }; + MsgRequestTopic req; + req.set_topic(topic); + req.set_data("data " + std::string(100, 'a')); + client.ClientStartWorker(onRecv, 2); + boost::timer::auto_cpu_timer timer; for (int i = 0; i < nreq; ++i) { - if (!client.ClientAsyncRequest(topic, "data " + std::to_string(i), 1000)) { + std::string msg_id; + if (!client.ClientAsyncRequest(req, msg_id)) { printf("client request failed\n"); ++count; } - // if (!client.SyncRequest(topic, "data " + std::to_string(i), reply, 1000)) { + // std::string proc_id; + // MsgRequestTopicReply reply; + // if (!client.ClientSyncRequest(req, proc_id, reply, 1000)) { // printf("client request failed\n"); // } - // ++count; + // ++count; } do { - std::this_thread::yield(); + std::this_thread::sleep_for(100ms); } while (count.load() < nreq); - client.StopAll(); + client.Stop(); printf("request %s %d done ", topic.c_str(), count.load()); }; @@ -248,12 +256,18 @@ auto Server = [&](const std::string &name, const std::vector<std::string> &topics) { DemoNode server(name, shm); - auto onData = [&](const std::string &topic, const std::string &data, std::string &reply) { + auto onDataSync = [&](const std::string &proc_id, const MsgRequestTopic &request, MsgRequestTopicReply &reply) { ++server_msg_count; - reply = topic + ':' + data; + reply.set_data(request.topic() + ':' + request.data()); return true; }; - server.ServerStart(onData); + auto onDataAsync = [&](void *src, std::string &proc_id, MsgRequestTopic &request) { + ++server_msg_count; + MsgRequestTopicReply reply; + reply.set_data(request.topic() + ':' + request.data()); + server.ServerSendReply(src, reply); + }; + server.ServerStart(onDataAsync); MsgTopicList rpc; for (auto &topic : topics) { @@ -266,15 +280,15 @@ } while (run) { - std::this_thread::yield(); + std::this_thread::sleep_for(100ms); } }; ThreadManager clients, servers; std::vector<Topic> topics = {"topic1", "topic2"}; servers.Launch(Server, "server", topics); - std::this_thread::sleep_for(100ms); + Sleep(100ms); for (auto &t : topics) { - clients.Launch(Client, t, 1000 * 1); + clients.Launch(Client, t, 1000 * 100 * 2); } clients.WaitAll(); printf("clients done, server replyed: %ld\n", server_msg_count.load()); @@ -282,6 +296,35 @@ servers.WaitAll(); } +BOOST_AUTO_TEST_CASE(HeartbeatTest) +{ + const std::string shm_name("ShmHeartbeat"); + ShmRemover auto_remove(shm_name); + SharedMemory shm(shm_name, 1024 * 1024 * 50); + + BHCenter center(shm); + center.Start(); + + { + + DemoNode node("demo_node", shm); + auto Check = [&]() { + bool r = node.Heartbeat(100); + printf("hearbeat ret : %s\n", r ? "ok" : "failed"); + }; + Check(); + for (int i = 0; i < 3; ++i) { + Sleep(1s); + Check(); + } + Sleep(4s); + for (int i = 0; i < 2; ++i) { + Sleep(1s); + Check(); + } + } + Sleep(8s); +} inline int MyMin(int a, int b) { printf("MyMin\n"); -- Gitblit v1.8.0