From 77a6c3512a44dfe6540dde71946e6484fe4f173f Mon Sep 17 00:00:00 2001
From: lichao <lichao@aiotlink.com>
Date: 星期一, 10 五月 2021 16:05:28 +0800
Subject: [PATCH] test lock code.

---
 utest/speed_test.cpp  |   65 +++++++++++++++++++++++---------
 src/robust.h          |   16 +++----
 src/shm_msg_queue.h   |    2 +
 src/shm_queue.h       |    6 +++
 utest/robust_test.cpp |   16 ++++++++
 src/robust.cpp        |   13 +++---
 src/shm_msg_queue.cpp |    1 
 7 files changed, 85 insertions(+), 34 deletions(-)

diff --git a/src/robust.cpp b/src/robust.cpp
index 08d2073..26d41b9 100644
--- a/src/robust.cpp
+++ b/src/robust.cpp
@@ -35,25 +35,24 @@
 
 bool FMutex::try_lock()
 {
-	if (flock(fd_, LOCK_EX | LOCK_NB) == 0) {
-		if (mtx_.try_lock()) {
+	if (mtx_.try_lock()) {
+		if (flock(fd_, LOCK_EX | LOCK_NB) == 0) {
 			return true;
 		} else {
-			flock(fd_, LOCK_UN);
+			mtx_.unlock();
 		}
 	}
 	return false;
 }
 void FMutex::lock()
 {
-	//Note: the lock order affects performance a lot,
-	// locking fd_ first is about 100 times faster than locking mtx_ first.
-	flock(fd_, LOCK_EX);
 	mtx_.lock();
+	flock(fd_, LOCK_EX);
 }
 void FMutex::unlock()
 {
-	mtx_.unlock();
 	flock(fd_, LOCK_UN);
+	mtx_.unlock();
 }
+
 } // namespace robust
\ No newline at end of file
diff --git a/src/robust.h b/src/robust.h
index d2d94e9..8657122 100644
--- a/src/robust.h
+++ b/src/robust.h
@@ -86,6 +86,8 @@
 class NullMutex
 {
 public:
+	template <class... T>
+	explicit NullMutex(T &&...t) {} // easy test.
 	bool try_lock() { return true; }
 	void lock() {}
 	void unlock() {}
@@ -113,8 +115,10 @@
 		mkdir(dir.c_str(), 0777);
 		return dir + "/fm_" + std::to_string(id);
 	}
-	static int Open(id_t id) { return open(GetPath(id).c_str(), O_CREAT | O_RDWR, 0666); }
+	static int Open(id_t id) { return open(GetPath(id).c_str(), O_CREAT | O_RDONLY, 0666); }
 	static int Close(int fd) { return close(fd); }
+	void FLock();
+	void FUnlock();
 	id_t id_;
 	int fd_;
 	std::mutex mtx_;
@@ -132,17 +136,11 @@
 {
 public:
 	SemMutex(key_t key) :
-	    key_(key), sem_id_(semget(key, 1, 0666 | IPC_CREAT))
+	    key_(key), sem_id_(semget(key, 1, 0666))
 	{
 		if (sem_id_ == -1) { throw "error create semaphore."; }
-		union semun init_val;
-		init_val.val = 1;
-		semctl(sem_id_, 0, SETVAL, init_val);
 	}
-	~SemMutex()
-	{
-		// semctl(sem_id_, 0, IPC_RMID, semun{});
-	}
+	~SemMutex() {}
 
 	bool try_lock()
 	{
diff --git a/src/shm_msg_queue.cpp b/src/shm_msg_queue.cpp
index bc5075f..b78c1a0 100644
--- a/src/shm_msg_queue.cpp
+++ b/src/shm_msg_queue.cpp
@@ -65,6 +65,7 @@
 	auto pos = imm.find(id);
 	if (pos == imm.end()) {
 		pos = imm.emplace(id, new Mutex(id)).first;
+		// pos = imm.emplace(id, new Mutex()).first;
 	}
 	return *pos->second;
 }
diff --git a/src/shm_msg_queue.h b/src/shm_msg_queue.h
index f8888f3..1970803 100644
--- a/src/shm_msg_queue.h
+++ b/src/shm_msg_queue.h
@@ -32,6 +32,8 @@
 	typedef Shmq::Data Queue;
 	typedef std::function<void()> OnSend;
 	typedef robust::FMutex Mutex;
+	// typedef robust::SemMutex Mutex;
+	// typedef robust::NullMutex Mutex;
 	typedef robust::Guard<Mutex> Guard;
 
 public:
diff --git a/src/shm_queue.h b/src/shm_queue.h
index 0041f16..5fd14e3 100644
--- a/src/shm_queue.h
+++ b/src/shm_queue.h
@@ -54,6 +54,7 @@
 	}
 	bool TryRead(D &d)
 	{
+		// bhome_shm::Guard lock(mutex_);
 		if (!queue_.empty()) {
 			d = queue_.front();
 			queue_.pop_front();
@@ -64,6 +65,7 @@
 	}
 	bool TryWrite(const D &d)
 	{
+		// bhome_shm::Guard lock(mutex_);
 		if (!queue_.full()) {
 			queue_.push_back(d);
 			return true;
@@ -74,12 +76,16 @@
 
 private:
 	Circular<D> queue_;
+	bhome_shm::Mutex mutex_;
 };
 
 template <int Power = 4>
 class SharedQ63
 {
 public:
+	template <class... T>
+	explicit SharedQ63(T &&...t) {} // easy testing
+
 	typedef int64_t Data;
 	bool Read(Data &d, const int timeout_ms)
 	{
diff --git a/utest/robust_test.cpp b/utest/robust_test.cpp
index 68c0e72..0645918 100644
--- a/utest/robust_test.cpp
+++ b/utest/robust_test.cpp
@@ -165,6 +165,22 @@
 
 BOOST_AUTO_TEST_CASE(MutexTest)
 {
+	{
+		int fd = open("/tmp/test_fmutex", O_CREAT | O_RDONLY, 0666);
+		flock(fd, LOCK_EX);
+		printf("lock 1");
+		Sleep(10s);
+		flock(fd, LOCK_EX);
+		printf("lock 2");
+		Sleep(10s);
+		flock(fd, LOCK_UN);
+		printf("un lock 2");
+		Sleep(10s);
+		flock(fd, LOCK_UN);
+		printf("un lock 1");
+		return;
+	}
+
 	// typedef robust::MFMutex RobustMutex;
 	typedef robust::SemMutex RobustMutex;
 
diff --git a/utest/speed_test.cpp b/utest/speed_test.cpp
index c512569..f8f54f5 100644
--- a/utest/speed_test.cpp
+++ b/utest/speed_test.cpp
@@ -15,22 +15,38 @@
  *
  * =====================================================================================
  */
+#include "robust.h"
 #include "util.h"
+
+using namespace robust;
 
 BOOST_AUTO_TEST_CASE(SpeedTest)
 {
-	const int mem_size = 1024 * 1024 * 50;
 	SharedMemory &shm = TestShm();
 	GlobalInit(shm);
+	auto InitSem = [](auto id) {
+		auto sem_id = semget(id, 1, 0666 | IPC_CREAT);
+		union semun init_val;
+		init_val.val = 1;
+		semctl(sem_id, 0, SETVAL, init_val);
+		return;
+	};
 
 	MQId id = ShmMsgQueue::NewId();
-	const int timeout = 1000;
-	const uint32_t data_size = 4000;
-	const std::string proc_id = "demo_proc";
+	InitSem(id);
 
+	const int timeout = 1000;
+	const uint32_t data_size = 1001;
+	const std::string proc_id = "demo_proc";
+	std::atomic<int64_t> nwrite(0);
+	std::atomic<int64_t> nread(0);
+
+	std::string str(data_size, 'a');
 	auto Writer = [&](int writer_id, uint64_t n) {
-		ShmMsgQueue mq(shm, 64);
-		std::string str(data_size, 'a');
+		MQId cli_id = ShmMsgQueue::NewId();
+		InitSem(cli_id);
+
+		ShmMsgQueue mq(cli_id, shm, 64);
 		MsgI msg;
 		MsgRequestTopic body;
 		body.set_topic("topic");
@@ -42,6 +58,7 @@
 
 		for (uint64_t i = 0; i < n; ++i) {
 			while (!mq.TrySend(id, msg)) {}
+			++nwrite;
 		}
 	};
 	auto Reader = [&](int reader_id, std::atomic<bool> *run, bool isfork) {
@@ -54,6 +71,7 @@
 			if (mq.TryRecv(msg)) {
 				DEFER1(msg.Release());
 				tm = now();
+				++nread;
 			} else if (isfork) {
 				if (now() > tm + 1s) {
 					exit(0); // for forked quit after 1s.
@@ -64,21 +82,24 @@
 	auto State = [&](std::atomic<bool> *run) {
 		auto init = shm.get_free_memory();
 		printf("shm init : %ld\n", init);
+		uint64_t last_read = 0;
 		while (*run) {
 			auto cur = shm.get_free_memory();
-			printf("shm used : %8ld/%ld\n", init - cur, init);
+			auto cur_read = nread.load();
+			printf("shm used : %8ld/%ld, write: %8ld, read: %8ld, speed: %8ld\n", init - cur, init, nwrite.load(), cur_read, cur_read - last_read);
+			last_read = cur_read;
 			std::this_thread::sleep_for(1s);
 		}
 	};
 
-	int nwriters[] = {1, 4, 16};
-	int nreaders[] = {1, 4};
+	int nwriters[] = {1, 10, 100};
+	int nreaders[] = {2};
+	const int64_t total_msg = 1000 * 100;
 
 	auto Test = [&](auto &www, auto &rrr, bool isfork) {
 		for (auto nreader : nreaders) {
 			for (auto nwriter : nwriters) {
-				const uint64_t nmsg = 1000 * 1000 * 10 / nwriter;
-				const uint64_t total_msg = nmsg * nwriter;
+				const uint64_t nmsg = total_msg / nwriter;
 				std::atomic<bool> run(true);
 				std::this_thread::sleep_for(10ms);
 				boost::timer::auto_cpu_timer timer;
@@ -100,16 +121,22 @@
 	std::atomic<bool> run(true);
 	ThreadManager state;
 	state.Launch(State, &run);
+	DEFER1(run.store(false););
+
 	// typedef ProcessManager Manager;
 	// typedef ThreadManager Manager;
 	// const bool isfork = IsSameType<Manager, ProcessManager>::value;
-	ProcessManager pw, pr;
-	printf("================ Testing process io: =======================================================\n");
-	Test(pw, pr, true);
-	ThreadManager tw, tr;
-	printf("---------------- Testing thread io:  -------------------------------------------------------\n");
-	Test(tw, tr, false);
-	run.store(false);
+
+	{
+		ThreadManager tw, tr;
+		printf("---------------- Testing thread io:  -------------------------------------------------------\n");
+		Test(tw, tr, false);
+	}
+	{
+		ProcessManager pw, pr;
+		printf("================ Testing process io: =======================================================\n");
+		Test(pw, pr, true);
+	}
 }
 
 // Send Recv Test
@@ -123,6 +150,8 @@
 	const std::string server_proc_id = "server_proc";
 
 	SharedMemory &shm = TestShm();
+	// shm.Remove();
+	// return;
 	GlobalInit(shm);
 
 	auto Avail = [&]() { return shm.get_free_memory(); };

--
Gitblit v1.8.0