From 9e5babf9db52e64bdae60137be7696e56241fca6 Mon Sep 17 00:00:00 2001 From: xingzilong <xingzilong@454eff88-639b-444f-9e54-f578c98de674> Date: 星期五, 18 八月 2017 18:12:17 +0800 Subject: [PATCH] H264 NALU解析 并在RTSPServer判断 --- RtspFace/PL_SensetimeFaceDetect.cpp | 208 +++++++++++++++------------------------------------ 1 files changed, 62 insertions(+), 146 deletions(-) diff --git a/RtspFace/PL_SensetimeFaceDetect.cpp b/RtspFace/PL_SensetimeFaceDetect.cpp index 85245fb..b6d633f 100644 --- a/RtspFace/PL_SensetimeFaceDetect.cpp +++ b/RtspFace/PL_SensetimeFaceDetect.cpp @@ -2,25 +2,19 @@ #include "MaterialBuffer.h" #include "logger.h" +//#include "SensetimeFaceAPIWrapper/src/FaceDBPool.h" +//#include "SensetimeFaceAPIWrapper/src/faceAPI.h" #include <opencv2/opencv.hpp> #include <cv_face.h> struct PL_SensetimeFaceDetect_Internal { - //uint8_t buffer[1920*1080*4]; - //size_t buffSize; - //size_t buffSizeMax; - MB_Frame lastFrame; - SensetimeFaceDetectConfig config; - bool payError; - - cv_handle_t handle_track; + PL_SensetimeFaceDetectConfig config; + SensetimeFaceDetectResult lastResult; PL_SensetimeFaceDetect_Internal() : - //buffSize(0), buffSizeMax(sizeof(buffer)), - lastFrame(), config(), payError(true), - handle_track(nullptr) + payError(true), config(), lastResult() { } @@ -30,15 +24,13 @@ void reset() { - //buffSize = 0; payError = true; - MB_Frame _lastFrame; - lastFrame = _lastFrame; - SensetimeFaceDetectConfig _config; + PL_SensetimeFaceDetectConfig _config; config = _config; - handle_track = nullptr; + SensetimeFaceDetectResult _lastResult; + lastResult = _lastResult; } }; @@ -62,156 +54,71 @@ PL_SensetimeFaceDetect_Internal* in = (PL_SensetimeFaceDetect_Internal*)internal; in->reset(); - SensetimeFaceDetectConfig* config = (SensetimeFaceDetectConfig*)args; - in->config = *config; - if (in->config.point_size == 21) - in->config.point_size_config = CV_DETECT_ENABLE_ALIGN_21; - else if (in->config.point_size == 106) - in->config.point_size_config = CV_DETECT_ENABLE_ALIGN_106; - else + if (args != nullptr) { - LOG(ERROR) << "alignment point size must be 21 or 106"; - return false; + PL_SensetimeFaceDetectConfig* config = (PL_SensetimeFaceDetectConfig*)args; + in->config = *config; } - // init handle - cv_result_t cv_result = cv_face_create_tracker(&(in->handle_track), nullptr, - in->config.point_size_config | CV_FACE_TRACKING_TWO_THREAD); - if (cv_result != CV_OK) - { - LOG(ERROR) << "cv_face_create_tracker failed, error code" << cv_result; - return false; - } - - int val = 0; - cv_result = cv_face_track_set_detect_face_cnt_limit(in->handle_track, in->config.detect_face_cnt_limit, &val); - if (cv_result != CV_OK) - { - LOG(ERROR) << "cv_face_track_set_detect_face_cnt_limit failed, error : " << cv_result; - return false; - } - else - LOG(ERROR) << "detect face count limit : " << val; - return true; } void PL_SensetimeFaceDetect::finit() { PL_SensetimeFaceDetect_Internal* in = (PL_SensetimeFaceDetect_Internal*)internal; - - // destroy track handle - cv_face_destroy_tracker(in->handle_track); - in->handle_track = nullptr; -} -int doFaceDetect(PL_SensetimeFaceDetect_Internal* in, - uint8_t* buffer, size_t width, size_t height, size_t stride, cv_pixel_format cvPixFmt) -{ - //resize(bgr_frame, bgr_frame, Size(frame_width, frame_height), 0, 0, INTER_LINEAR); - - int face_count = 0; - cv_result_t cv_result = CV_OK; - cv_face_t* p_face = nullptr; - - // realtime track - cv_result = cv_face_track(in->handle_track, buffer, cvPixFmt, - width, height, stride, - CV_FACE_UP, &p_face, &face_count); - if (cv_result != CV_OK) - { - LOG(ERROR) << "cv_face_track failed, error : " << cv_result; - cv_face_release_tracker_result(p_face, face_count); - return -1; - } - - // draw the video - cv::Mat yuvMat(cv::Size(1920,1080), CV_8UC3, buffer);//#todo - cv::Mat yMat(cv::Size(1920,1080), CV_8UC1, buffer); - for (int i = 0; i < face_count; i++) - { - LOGP(DEBUG, "face: %d-----[%d, %d, %d, %d]-----id: %d\n", i, - p_face[i].rect.left, p_face[i].rect.top, - p_face[i].rect.right, p_face[i].rect.bottom, p_face[i].ID); - - LOGP(DEBUG, "face pose: [yaw: %.2f, pitch: %.2f, roll: %.2f, eye distance: %.2f]\n", - p_face[i].yaw, - p_face[i].pitch, p_face[i].roll, p_face[i].eye_dist); - - cv::Scalar scalar_color = CV_RGB(p_face[i].ID * 53 % 256, - p_face[i].ID * 93 % 256, - p_face[i].ID * 143 % 256); - - //cv::rectangle(yMat, cv::Point2f(0, 0), cv::Point2f(50, 50), scalar_color, 2); - //cv::rectangle(yMat, cv::Point2f(500, 500), cv::Point2f(550, 550), scalar_color, 2); - - cv::rectangle(yMat, cv::Point2f(static_cast<float>(p_face[i].rect.left), - static_cast<float>(p_face[i].rect.top)), - cv::Point2f(static_cast<float>(p_face[i].rect.right), - static_cast<float>(p_face[i].rect.bottom)), scalar_color, 2); - - for (int j = 0; j < p_face[i].points_count; j++) - { - cv::circle(yMat, cv::Point2f(p_face[i].points_array[j].x, - p_face[i].points_array[j].y), 1, cv::Scalar(255, 255, 255)); - } - } - - //if (face_count > 0) - //{ - // static size_t f=0; - // char fname[50]; - // sprintf(fname, "face-%u.yuv420", ++f); - // FILE * pFile = fopen (fname,"wb"); - // fwrite (yuvMat.data , sizeof(char), 1920*1080*1.5, pFile); - // printf("write face file %s\n", fname); - // fclose(pFile); - //} - - // release the memory of face - cv_face_release_tracker_result(p_face, face_count); - - return face_count; } bool PL_SensetimeFaceDetect::pay(const PipeMaterial& pm) { PL_SensetimeFaceDetect_Internal* in = (PL_SensetimeFaceDetect_Internal*)internal; - - if (pm.type != PipeMaterial::PMT_FRAME) + + in->payError = true; { - LOG(ERROR) << "PL_H264Encoder::pay only support PMT_FRAME"; + SensetimeFaceDetectResult _lastResult; + in->lastResult = _lastResult; + } + + if (pm.type != PipeMaterial::PMT_FRAME || !(in->config.payWithDbFrame)) + { + LOG_ERROR << "Only support PMT_FRAME (SensetimeFaceDetectDbFrame)" << std::endl; return false; } if (pm.buffer == nullptr) return false; - MB_Frame* frame = (MB_Frame*)pm.buffer; + SensetimeFaceDetectDbFrame* frame = (SensetimeFaceDetectDbFrame*)pm.buffer; if (frame->type != MB_Frame::MBFT_YUV420) { - LOG(ERROR) << "PL_H264Encoder::pay only support MBFT_YUV420"; + LOG_ERROR << "Only support MBFT_YUV420" << std::endl; return false; } - int face_count = doFaceDetect(in, (uint8_t*)frame->buffer, 1920, 1080, 1920, CV_PIX_FMT_YUV420P);//#todo - if (face_count < 0) - { - in->payError = true; - return false; - } - else - in->payError = false; + //FaceDB* _faceDB = (FaceDB*)frame->_faceDB; + //if (_faceDB == nullptr) + //{ + // LOG_ERROR << "FaceDB is null" << std::endl; + // return false; + //} - //in->buffer readly - - in->lastFrame.type = MB_Frame::MBFT_YUV420; - in->lastFrame.buffer = frame->buffer;//#todo should copy - in->lastFrame.buffSize = frame->buffSize; - in->lastFrame.width = frame->width; - in->lastFrame.height = frame->height; - in->lastFrame.pts = frame->pts; - + const size_t expectedYUVSize = frame->width * frame->height * 1.5; + if (frame->buffSize < expectedYUVSize) + { + LOG_WARN << "image size not ok" << std::endl; + return false; + } + + cv::Mat yuvMat(cv::Size(frame->width,frame->height), CV_8UC3, frame->buffer); + //cv_feature_t* feature = _faceDB->extract_feature(yuvMat); + //in->lastResult.st_id = _faceDB->search_db(feature); + + //in->lastResult.st_id = _faceDB->do_reasch(yuvMat); + //#todo release feature + + in->lastResult.school_id = frame->school_id; + + in->payError = false; return true; } @@ -219,13 +126,22 @@ { PL_SensetimeFaceDetect_Internal* in = (PL_SensetimeFaceDetect_Internal*)internal; - if (!in->payError) - { - pm.type = PipeMaterial::PMT_FRAME; - pm.buffer = &(in->lastFrame); - pm.buffSize = 0; - pm.former = this; - } pm.former = this; - return !in->payError; + + if (in->payError) + { + return false; + } + + if (! in->config.resultStructOnly) + { + LOG_ERROR << "Only support resultStructOnly" << std::endl; + return false; + } + + pm.type = PipeMaterial::PMT_BYTES; + pm.buffer = &(in->lastResult); + pm.buffSize = sizeof(SensetimeFaceDetectResult); + + return true; } -- Gitblit v1.8.0