From 77a6c3512a44dfe6540dde71946e6484fe4f173f Mon Sep 17 00:00:00 2001
From: lichao <lichao@aiotlink.com>
Date: 星期一, 10 五月 2021 16:05:28 +0800
Subject: [PATCH] test lock code.
---
utest/speed_test.cpp | 65 +++++++++++++++++++++++---------
src/robust.h | 16 +++----
src/shm_msg_queue.h | 2 +
src/shm_queue.h | 6 +++
utest/robust_test.cpp | 16 ++++++++
src/robust.cpp | 13 +++---
src/shm_msg_queue.cpp | 1
7 files changed, 85 insertions(+), 34 deletions(-)
diff --git a/src/robust.cpp b/src/robust.cpp
index 08d2073..26d41b9 100644
--- a/src/robust.cpp
+++ b/src/robust.cpp
@@ -35,25 +35,24 @@
bool FMutex::try_lock()
{
- if (flock(fd_, LOCK_EX | LOCK_NB) == 0) {
- if (mtx_.try_lock()) {
+ if (mtx_.try_lock()) {
+ if (flock(fd_, LOCK_EX | LOCK_NB) == 0) {
return true;
} else {
- flock(fd_, LOCK_UN);
+ mtx_.unlock();
}
}
return false;
}
void FMutex::lock()
{
- //Note: the lock order affects performance a lot,
- // locking fd_ first is about 100 times faster than locking mtx_ first.
- flock(fd_, LOCK_EX);
mtx_.lock();
+ flock(fd_, LOCK_EX);
}
void FMutex::unlock()
{
- mtx_.unlock();
flock(fd_, LOCK_UN);
+ mtx_.unlock();
}
+
} // namespace robust
\ No newline at end of file
diff --git a/src/robust.h b/src/robust.h
index d2d94e9..8657122 100644
--- a/src/robust.h
+++ b/src/robust.h
@@ -86,6 +86,8 @@
class NullMutex
{
public:
+ template <class... T>
+ explicit NullMutex(T &&...t) {} // easy test.
bool try_lock() { return true; }
void lock() {}
void unlock() {}
@@ -113,8 +115,10 @@
mkdir(dir.c_str(), 0777);
return dir + "/fm_" + std::to_string(id);
}
- static int Open(id_t id) { return open(GetPath(id).c_str(), O_CREAT | O_RDWR, 0666); }
+ static int Open(id_t id) { return open(GetPath(id).c_str(), O_CREAT | O_RDONLY, 0666); }
static int Close(int fd) { return close(fd); }
+ void FLock();
+ void FUnlock();
id_t id_;
int fd_;
std::mutex mtx_;
@@ -132,17 +136,11 @@
{
public:
SemMutex(key_t key) :
- key_(key), sem_id_(semget(key, 1, 0666 | IPC_CREAT))
+ key_(key), sem_id_(semget(key, 1, 0666))
{
if (sem_id_ == -1) { throw "error create semaphore."; }
- union semun init_val;
- init_val.val = 1;
- semctl(sem_id_, 0, SETVAL, init_val);
}
- ~SemMutex()
- {
- // semctl(sem_id_, 0, IPC_RMID, semun{});
- }
+ ~SemMutex() {}
bool try_lock()
{
diff --git a/src/shm_msg_queue.cpp b/src/shm_msg_queue.cpp
index bc5075f..b78c1a0 100644
--- a/src/shm_msg_queue.cpp
+++ b/src/shm_msg_queue.cpp
@@ -65,6 +65,7 @@
auto pos = imm.find(id);
if (pos == imm.end()) {
pos = imm.emplace(id, new Mutex(id)).first;
+ // pos = imm.emplace(id, new Mutex()).first;
}
return *pos->second;
}
diff --git a/src/shm_msg_queue.h b/src/shm_msg_queue.h
index f8888f3..1970803 100644
--- a/src/shm_msg_queue.h
+++ b/src/shm_msg_queue.h
@@ -32,6 +32,8 @@
typedef Shmq::Data Queue;
typedef std::function<void()> OnSend;
typedef robust::FMutex Mutex;
+ // typedef robust::SemMutex Mutex;
+ // typedef robust::NullMutex Mutex;
typedef robust::Guard<Mutex> Guard;
public:
diff --git a/src/shm_queue.h b/src/shm_queue.h
index 0041f16..5fd14e3 100644
--- a/src/shm_queue.h
+++ b/src/shm_queue.h
@@ -54,6 +54,7 @@
}
bool TryRead(D &d)
{
+ // bhome_shm::Guard lock(mutex_);
if (!queue_.empty()) {
d = queue_.front();
queue_.pop_front();
@@ -64,6 +65,7 @@
}
bool TryWrite(const D &d)
{
+ // bhome_shm::Guard lock(mutex_);
if (!queue_.full()) {
queue_.push_back(d);
return true;
@@ -74,12 +76,16 @@
private:
Circular<D> queue_;
+ bhome_shm::Mutex mutex_;
};
template <int Power = 4>
class SharedQ63
{
public:
+ template <class... T>
+ explicit SharedQ63(T &&...t) {} // easy testing
+
typedef int64_t Data;
bool Read(Data &d, const int timeout_ms)
{
diff --git a/utest/robust_test.cpp b/utest/robust_test.cpp
index 68c0e72..0645918 100644
--- a/utest/robust_test.cpp
+++ b/utest/robust_test.cpp
@@ -165,6 +165,22 @@
BOOST_AUTO_TEST_CASE(MutexTest)
{
+ {
+ int fd = open("/tmp/test_fmutex", O_CREAT | O_RDONLY, 0666);
+ flock(fd, LOCK_EX);
+ printf("lock 1");
+ Sleep(10s);
+ flock(fd, LOCK_EX);
+ printf("lock 2");
+ Sleep(10s);
+ flock(fd, LOCK_UN);
+ printf("un lock 2");
+ Sleep(10s);
+ flock(fd, LOCK_UN);
+ printf("un lock 1");
+ return;
+ }
+
// typedef robust::MFMutex RobustMutex;
typedef robust::SemMutex RobustMutex;
diff --git a/utest/speed_test.cpp b/utest/speed_test.cpp
index c512569..f8f54f5 100644
--- a/utest/speed_test.cpp
+++ b/utest/speed_test.cpp
@@ -15,22 +15,38 @@
*
* =====================================================================================
*/
+#include "robust.h"
#include "util.h"
+
+using namespace robust;
BOOST_AUTO_TEST_CASE(SpeedTest)
{
- const int mem_size = 1024 * 1024 * 50;
SharedMemory &shm = TestShm();
GlobalInit(shm);
+ auto InitSem = [](auto id) {
+ auto sem_id = semget(id, 1, 0666 | IPC_CREAT);
+ union semun init_val;
+ init_val.val = 1;
+ semctl(sem_id, 0, SETVAL, init_val);
+ return;
+ };
MQId id = ShmMsgQueue::NewId();
- const int timeout = 1000;
- const uint32_t data_size = 4000;
- const std::string proc_id = "demo_proc";
+ InitSem(id);
+ const int timeout = 1000;
+ const uint32_t data_size = 1001;
+ const std::string proc_id = "demo_proc";
+ std::atomic<int64_t> nwrite(0);
+ std::atomic<int64_t> nread(0);
+
+ std::string str(data_size, 'a');
auto Writer = [&](int writer_id, uint64_t n) {
- ShmMsgQueue mq(shm, 64);
- std::string str(data_size, 'a');
+ MQId cli_id = ShmMsgQueue::NewId();
+ InitSem(cli_id);
+
+ ShmMsgQueue mq(cli_id, shm, 64);
MsgI msg;
MsgRequestTopic body;
body.set_topic("topic");
@@ -42,6 +58,7 @@
for (uint64_t i = 0; i < n; ++i) {
while (!mq.TrySend(id, msg)) {}
+ ++nwrite;
}
};
auto Reader = [&](int reader_id, std::atomic<bool> *run, bool isfork) {
@@ -54,6 +71,7 @@
if (mq.TryRecv(msg)) {
DEFER1(msg.Release());
tm = now();
+ ++nread;
} else if (isfork) {
if (now() > tm + 1s) {
exit(0); // for forked quit after 1s.
@@ -64,21 +82,24 @@
auto State = [&](std::atomic<bool> *run) {
auto init = shm.get_free_memory();
printf("shm init : %ld\n", init);
+ uint64_t last_read = 0;
while (*run) {
auto cur = shm.get_free_memory();
- printf("shm used : %8ld/%ld\n", init - cur, init);
+ auto cur_read = nread.load();
+ printf("shm used : %8ld/%ld, write: %8ld, read: %8ld, speed: %8ld\n", init - cur, init, nwrite.load(), cur_read, cur_read - last_read);
+ last_read = cur_read;
std::this_thread::sleep_for(1s);
}
};
- int nwriters[] = {1, 4, 16};
- int nreaders[] = {1, 4};
+ int nwriters[] = {1, 10, 100};
+ int nreaders[] = {2};
+ const int64_t total_msg = 1000 * 100;
auto Test = [&](auto &www, auto &rrr, bool isfork) {
for (auto nreader : nreaders) {
for (auto nwriter : nwriters) {
- const uint64_t nmsg = 1000 * 1000 * 10 / nwriter;
- const uint64_t total_msg = nmsg * nwriter;
+ const uint64_t nmsg = total_msg / nwriter;
std::atomic<bool> run(true);
std::this_thread::sleep_for(10ms);
boost::timer::auto_cpu_timer timer;
@@ -100,16 +121,22 @@
std::atomic<bool> run(true);
ThreadManager state;
state.Launch(State, &run);
+ DEFER1(run.store(false););
+
// typedef ProcessManager Manager;
// typedef ThreadManager Manager;
// const bool isfork = IsSameType<Manager, ProcessManager>::value;
- ProcessManager pw, pr;
- printf("================ Testing process io: =======================================================\n");
- Test(pw, pr, true);
- ThreadManager tw, tr;
- printf("---------------- Testing thread io: -------------------------------------------------------\n");
- Test(tw, tr, false);
- run.store(false);
+
+ {
+ ThreadManager tw, tr;
+ printf("---------------- Testing thread io: -------------------------------------------------------\n");
+ Test(tw, tr, false);
+ }
+ {
+ ProcessManager pw, pr;
+ printf("================ Testing process io: =======================================================\n");
+ Test(pw, pr, true);
+ }
}
// Send Recv Test
@@ -123,6 +150,8 @@
const std::string server_proc_id = "server_proc";
SharedMemory &shm = TestShm();
+ // shm.Remove();
+ // return;
GlobalInit(shm);
auto Avail = [&]() { return shm.get_free_memory(); };
--
Gitblit v1.8.0