From 18a05d269516a5e33d8460291c2f93e73d95adce Mon Sep 17 00:00:00 2001 From: zhangmeng <775834166@qq.com> Date: 星期二, 26 十二月 2023 10:45:31 +0800 Subject: [PATCH] GetYUV format is NV12 --- csrc/wrapper.cpp | 412 ++++++++++++++++++++++++++++++---------------------------- 1 files changed, 211 insertions(+), 201 deletions(-) diff --git a/csrc/wrapper.cpp b/csrc/wrapper.cpp index 7149cbf..a84fd19 100644 --- a/csrc/wrapper.cpp +++ b/csrc/wrapper.cpp @@ -1,7 +1,9 @@ #include "wrapper.hpp" #include <thread> +#include <atomic> #include <unistd.h> +#include <sys/time.h> extern "C"{ #include <libavformat/avformat.h> @@ -23,8 +25,7 @@ #include "worker/stream.hpp" #include "worker/decoder.hpp" #include "worker/rec.hpp" - -#include "CUDALERP.h" +#include "common.hpp" using namespace logif; using namespace ffwrapper; @@ -45,13 +46,20 @@ ,audio_(false) ,gb_(0) ,cpu_(0) + ,devid_(-1) ,run_dec_(false) + ,run_stream_(true) + ,run_rec_(false) + ,work_start(false) ,thread_(nullptr) ,stop_stream_(false) ,stream_(nullptr) ,decoder_(nullptr) ,rec_(new rec) ,logit_(false) + ,fps_(25) + ,file_finished_(false) + ,min_duration_(0) { makeTheWorld(); } @@ -61,13 +69,20 @@ ,audio_(false) ,gb_(0) ,cpu_(0) + ,devid_(-1) ,run_dec_(false) + ,run_stream_(true) + ,run_rec_(false) + ,work_start(false) ,thread_(nullptr) ,stop_stream_(false) ,stream_(nullptr) ,decoder_(nullptr) ,rec_(new rec) ,logit_(true) + ,fps_(25) + ,file_finished_(false) + ,min_duration_(0) { makeTheWorld(); logif::CreateLogger(logfile, true); @@ -90,6 +105,8 @@ } if (logit_) logif::DestroyLogger(); + + printf("wrapper release\n"); } std::unique_ptr<ffwrapper::FormatIn> Wrapper::init_reader(const char* input){ @@ -97,27 +114,26 @@ VideoProp prop; prop.url_ = input; prop.rtsp_tcp_ = true; - prop.gpu_acc_ = !cpu_; + prop.gpu_acc_ = false; + prop.gpu_index_ = devid_; - std::unique_ptr<FormatIn> in(new FormatIn(prop.gpuAccl())); - AVDictionary *avdic = prop.optsFormat(); + std::unique_ptr<FormatIn> in(nullptr); int flag = -1; + AVDictionary* avdic = NULL; if (gb_){ - flag = in->openGb28181(input, NULL); + in.reset(new FormatInGB(prop)); }else{ + in.reset(new FormatIn(prop, prop.gpuAccl())); + avdic = prop.optsFormat(); + } + + if(avdic){ flag = in->open(input, &avdic); + av_dict_free(&avdic); + }else{ + flag = in->open(input, NULL); } - if(avdic){ - av_dict_free(&avdic); - } - if(flag == 0){ - if(!in->findStreamInfo(NULL)){ - logIt("can't find video stream\n"); - return nullptr; - } - - return in; - } + if(flag == 0) return in; return nullptr; } @@ -149,7 +165,7 @@ stream_ = new stream(in, 3 * in->getFPS()); // stream_->AudioSwitch(audio_); - + decoder_ = new decoder(in); rec_->Load(in); @@ -159,15 +175,15 @@ } } - int Wrapper::run_worker(ffwrapper::FormatIn *in, std::shared_ptr<ffwrapper::CodedData> data, int64_t &id){ + int Wrapper::run_worker(ffwrapper::FormatIn *in, const CPacket &pkt){ if (gb_){ - AVPacket &pkt = data->getAVPacket(); - pkt.pts = pkt.dts = AV_NOPTS_VALUE; + AVPacket &p = pkt.data->getAVPacket(); + // p.pts = p.dts = AV_NOPTS_VALUE; } int flag = 0; - if (stream_) stream_->SetPacket(data, id); - if (decoder_ && run_dec_) flag = decoder_->SetFrame(data, id); - if (rec_->Loaded()) rec_->SetPacket(data, id); + if (run_stream_ && stream_) stream_->SetPacket(pkt); + if (run_dec_ && decoder_) flag = decoder_->SetFrame(pkt); + if (run_rec_ && rec_->Loaded()) rec_->SetPacket(pkt); return flag; } @@ -180,90 +196,189 @@ void Wrapper::run_stream_thread(){ + int64_t file_rebootime = 0; + int64_t file_frame = 0; + + using namespace std; + // const string gb_suffix[] = {"/StreamType=2", "/StreamType=3", ""}; + // const string gb_suffix[] = {"/StreamType=2", ""}; + const string gb_suffix[] = {""}; + const size_t gb_size = sizeof(gb_suffix) / sizeof(string); + int gs_idx = 0; + string url; + url.reserve(input_url_.size()*2); + url = input_url_; + work_start = false; while(!stop_stream_.load()){ - auto in = init_reader(input_url_.c_str()); + if (gb_){ + if (input_url_.find("/StreamType=") == string::npos) + url = input_url_ + gb_suffix[gs_idx]; + logIt("======>>input real url %s\n", url.c_str()); + gs_idx = (gs_idx + 1) % gb_size; + } + auto in = init_reader(url.c_str()); if (!in) { - logIt("ERROR: init_reader! url: %s\n", input_url_.c_str()); - sleep(2); + work_start = false; + logIt("ERROR: init_reader! url: %s\n", url.c_str()); + usleep(126000); + // for(int i = 0; i < 10; i++){ + // if (stop_stream_.load()) break; + if(gb_) usleep(2617000); + // } continue; } + + fps_ = in->getFPS(); - int wTime = 1000000.0 / in->getFPS() ; + if (fps_ < 1 || fps_ > 200){ + fps_ = 25; + } + + int wTime = 1000000.0 / fps_ ; wTime >>= 1; - logIt("WAIT TIME PER FRAME: %d", wTime); init_worker(in.get()); + work_start = true; int64_t id = gb_ ? 0 : -1; + int64_t v_id = id; + int64_t a_id = id; + + bool exist = access(input_url_.c_str(), 0) == 0; + if (exist){ + wTime += wTime/2; + file_rebootime++; + } + + logIt("WAIT TIME PER FRAME: %d", wTime); while(!stop_stream_.load()){ auto data(std::make_shared<CodedData>()); - if (in->readPacket(&data->getAVPacket()) != 0){ + auto ret = in->readPacket(&data->getAVPacket()); + if (ret > 0){ + sleep(2); + continue; + } + + if (ret < 0){ logIt("read packet error, id: %lld", id); break; } - + // 闈為煶瑙嗛 if (in->notVideoAudio(&data->getAVPacket())){ continue; } - + + // 闈炲浗鏍囪烦杩囩涓�甯�,娴嬭瘯绗竴甯ф湁闂 if (!gb_ && id < 0){ - id++; + id++; v_id++; a_id++; continue; } + CPacket pkt{data, v_id, a_id, id}; // decode error - if (run_worker(in.get(), data, id) == -1){ + if (run_worker(in.get(), pkt) == -1){ break; } - usleep(wTime); + + if (in->isVideoPkt(&data->getAVPacket())){ + v_id++; + }else{ + a_id++; + } id++; + + //鏈湴鏂囦欢澶揩sleep涓�涓� + if (exist && in->isVideoPkt(&data->getAVPacket())){ + // 璁板綍瑙g爜鐨勬枃浠舵湁澶氬皯甯� + file_frame++; + usleep(wTime); + } } deinit_worker(); + if (exist) { + // 涓夋涓�甯ч兘娌¤В鍑烘潵锛岄��鍑� + if (file_frame == 0 && file_rebootime < 3){ + logIt("libcffmpeg.so-> run %lld time, current frames %lld", file_rebootime, file_frame); + continue; + } + // 鏈�灏忛渶瑕佽В鍑哄灏戝抚 + if (file_frame < min_duration_ * fps_){ + logIt("libcffmpeg.so-> run %lld time, current frames %lld, mind %d, fps: %d", + file_rebootime, file_frame, min_duration_, fps_); + continue; + } + logIt("libcffmpeg.so-> local file %s run %lld time, all frames %lld, mind %d, fps: %d", + input_url_.c_str(), file_rebootime, file_frame, min_duration_, fps_); + file_finished_ = true; + break; + } } } - void Wrapper::BuildRecorder(const char* id, const char *output, const int mindur, const int maxdur, const bool audio){ + void Wrapper::SetRecMinCacheTime(const int mind, const int maxd){ + min_duration_ = mind; + } + + void Wrapper::OpenRecorder(){ + run_rec_ = true; + } + + void Wrapper::BuildRecorder(const char* id, const char *output, const int64_t &fid, const int mindur, const int maxdur, const bool audio){ bool a = audio; if (gb_) a = false; - + if (rec_->Loaded()){ - rec_->NewRec(id, output, mindur, maxdur, a); + rec_->NewRec(id, output, fid, mindur, maxdur, a); }else{ std::string rid(id), dir(output); fn_rec_lazy_ = - [=]{rec_->NewRec(rid.c_str(), dir.c_str(), mindur, maxdur, a);}; + [=]{rec_->NewRec(rid.c_str(), dir.c_str(), fid, mindur, maxdur, a);}; } } - int Wrapper::FireRecorder(const char* sid,const int64_t &id){ if (rec_->Loaded()){ rec_->FireRecSignal(sid, id); } + return 0; } - void Wrapper::GetInfoRecorder(std::string &recID, int &index, std::string &path){ - if (rec_){ + int Wrapper::GetInfoRecorder(std::string &recID, int &index, std::string &path){ + if (work_start && rec_){ rec_->GetRecInfo(recID, index, path); } + return 0; } ////////decoder - void Wrapper::BuildDecoder(){ + void Wrapper::OpenDecoder(){ run_dec_ = true; } - void Wrapper::GetPicDecoder(unsigned char **data, int *w, int *h, int *format, int *length, int64_t *id){ + int Wrapper::GetPicDecoder(unsigned char **data, int *w, int *h, int *format, int *length, int64_t *id){ + if (file_finished_){ + return -1; + } if (decoder_){ decoder_->GetFrame(data, w, h, format, length, id); + }else { + if (work_start){ + for(int i = 0; i < 6; i++)this_thread::sleep_for(chrono::seconds{1}); + } } + return 0; } - void Wrapper::GetPacket(unsigned char **pktData, int *size, int *key){ - if (stream_){ + void Wrapper::CloseStream(){ + run_stream_ = false; + } + + int Wrapper::GetPacket(unsigned char **pktData, int *size, int *key){ + if (work_start && stream_){ stream_->GetPacket(pktData, size, key); } + return 0; } } // end class wrapper @@ -275,69 +390,6 @@ // return val: -1 open error; -2, find stream error; -3, converter create namespace cffmpeg_wrap{ // start test functions - uint8_t* Decode(const char *file, const int gb, int *w, int *h){ - VideoProp prop; - prop.url_ = file; - prop.gpu_acc_ = false; - - std::unique_ptr<FormatIn> in(new FormatIn(prop.gpuAccl())); - int flag = -1; - if (gb){ - flag = in->openGb28181(file, NULL); - }else{ - flag = in->open(file, NULL); - } - - std::unique_ptr<cvbridge> bridge_(nullptr); - - if(flag == 0){ - if(!in->findStreamInfo(NULL)){ - logIt("yolo can't find video stream\n"); - *w = *h = -2; - return NULL; - } - auto flag = in->openCodec(NULL); - if(flag){ - auto dec_ctx = in->getCodecContext(); - - AVPixelFormat pix_fmt = AV_PIX_FMT_BGR24; - bridge_.reset(new cvbridge( - dec_ctx->width, dec_ctx->height, dec_ctx->pix_fmt, - dec_ctx->width, dec_ctx->height, pix_fmt, SWS_BICUBIC)); - - }else{ - logIt("FormatIn openCodec Failed!"); - *w = *h = -3; - return NULL; - } - }else{ - logIt("open %s error", file); - *w = *h = -1; - return NULL; - } - - uint8_t *pic = NULL; - *w = *h = 0; - - int tryTime = 0; - while (tryTime++ < 100){ - - auto data(std::make_shared<CodedData>()); - if (in->readPacket(&data->getAVPacket()) == 0){ - - auto frame(std::make_shared<FrameData>()); - AVFrame *frm = frame->getAVFrame(); - if(in->decode(frm, &data->getAVPacket()) == 0){ - *w = frm->width; - *h = frm->height; - pic = bridge_->convert2Data(frm); - break; - } - } - } - - return pic; - } /////// for encoder typedef struct _PicEncoder{ FormatOut *enc; @@ -346,11 +398,13 @@ int fps; int br; int gi; + int pix_fmt; int flag; cvbridge *bridge; } PicEncoder; - void *CreateEncoder(const int w, const int h, const int fps, const int br, const int scale_flag, const int gi){ + void *CreateEncoder(const int w, const int h, const int fps, const int br, + const int pix_fmt, const int scale_flag, const int gi){ PicEncoder *e = (PicEncoder*)malloc(sizeof(PicEncoder)); e->enc = NULL; @@ -359,6 +413,7 @@ e->fps = fps; e->br = br; e->gi = gi; + e->pix_fmt = pix_fmt; e->flag = scale_flag; e->bridge = NULL; @@ -392,14 +447,16 @@ PicEncoder *e = (PicEncoder*)hdl; auto ctx = e->enc->getCodecContext(); - AVPixelFormat pix_fmt = AV_PIX_FMT_BGR24; if (e->bridge == NULL){ e->bridge = new cvbridge( - w, h, AV_PIX_FMT_BGR24, + w, h, e->pix_fmt, e->w, e->h, ctx->pix_fmt, e->flag); } - AVFrame *frame = cvbridge::fillFrame(in, w, h, pix_fmt); + AVFrame *bgr_frame = cvbridge::fillFrame(in, w, h, e->pix_fmt); + AVFrame *frame = e->bridge->convert2Frame(bgr_frame); + av_frame_free(&bgr_frame); + AVPacket *pkt = av_packet_alloc(); auto flag = e->enc->encode(pkt, frame); @@ -427,96 +484,49 @@ return flag; } -/////////////////////////////////////////////////////////// - typedef struct _conv - { - int srcW; - int srcH; - int srcF; - int dstW; - int dstH; - cvbridge *b; - }Conv; - - void *CreateConvertor(const int srcW, const int srcH, const int srcFormat, - const int dstW, const int dstH, const int dstFormat, const int flag){ - - auto bridge = new cvbridge( - srcW, srcH, srcFormat, - dstW, dstH, dstFormat, flag); - if (!bridge) return NULL; +#ifdef GB28181 + int GetGb28181Pic(const char *rtspUrl, char *retData, int *retDataLen, const int tt){ + int ret = 0; + std::atomic_bool add_camera{false}; - Conv *c = (Conv*)malloc(sizeof(Conv)); - c->b = bridge; - c->dstW = dstW; - c->dstH = dstH; - c->srcW = srcW; - c->srcH = srcH; - c->srcF = srcFormat; + std::thread t([&]{ + auto handle_gb28181 = new GB28181API; + std::string ru(rtspUrl); + if(handle_gb28181->addCamera(ru) == -1){ + delete(handle_gb28181); + logIt("do addCamera Error\n"); + ret = -2; + return; + } + add_camera.store(true); + std::this_thread::sleep_for(std::chrono::seconds(1)); + int retLen = handle_gb28181->capturePic(handle_gb28181, retData, retDataLen, tt); + if(retLen == 0){ + logIt("do capturePic failed:%d"); + ret = -1; + } - return c; + handle_gb28181->deleteCamera(); + delete handle_gb28181; + }); + + std::this_thread::sleep_for(std::chrono::seconds(1)); + + int ttt = tt; + if (ttt < 6) ttt = 6; + + int i = -1; + while (true) { + std::this_thread::sleep_for(std::chrono::seconds(1)); + if (add_camera.load()) { + t.join(); + return ret; + } + if (i++ > ttt) break; + } + t.detach(); + return -1; } - - uint8_t *Convert(void *h, uint8_t *src){ - Conv *c = (Conv*)h; - - auto b = c->b; - - AVFrame *tmp_frm = cvbridge::fillFrame(src, c->srcW, c->srcH, c->srcF); - if (!tmp_frm) return NULL; - - unsigned char *picData = b->convert2Data(tmp_frm); - - av_frame_free(&tmp_frm); - - return picData; - } - - void DestoryConvertor(void *h){ - Conv *c = (Conv*)h; - delete c->b; - free(c); - } - - - uint8_t* ConvertYUV2BGR(uint8_t *src, const int w, const int h, const int dst_w, const int dst_h, int *length){ - return NULL; - - // int oldw = w, oldh = h, neww = dst_w, newh = dst_h; - // // setting cache and shared modes - // cudaDeviceSetCacheConfig(cudaFuncCachePreferL1); - // cudaDeviceSetSharedMemConfig(cudaSharedMemBankSizeFourByte); - - // // allocating and transferring image and binding to texture object - // cudaChannelFormatDesc chandesc_img = cudaCreateChannelDesc(8, 0, 0, 0, cudaChannelFormatKindUnsigned); - // cudaArray* d_img_arr; - // cudaMallocArray(&d_img_arr, &chandesc_img, oldw, oldh, cudaArrayTextureGather); - // cudaMemcpyToArray(d_img_arr, 0, 0, image, oldh * oldw, cudaMemcpyHostToDevice); - // struct cudaResourceDesc resdesc_img; - // memset(&resdesc_img, 0, sizeof(resdesc_img)); - // resdesc_img.resType = cudaResourceTypeArray; - // resdesc_img.res.array.array = d_img_arr; - // struct cudaTextureDesc texdesc_img; - // memset(&texdesc_img, 0, sizeof(texdesc_img)); - // texdesc_img.addressMode[0] = cudaAddressModeClamp; - // texdesc_img.addressMode[1] = cudaAddressModeClamp; - // texdesc_img.readMode = cudaReadModeNormalizedFloat; - // texdesc_img.filterMode = cudaFilterModePoint; - // texdesc_img.normalizedCoords = 0; - // cudaTextureObject_t d_img_tex = 0; - // cudaCreateTextureObject(&d_img_tex, &resdesc_img, &texdesc_img, nullptr); - - // uint8_t* d_out = nullptr; - // cudaMalloc(&d_out, total); - - // for (int i = 0; i < warmups; ++i) CUDALERP(d_img_tex, oldw, oldh, d_out, neww, newh); - // auto start = high_resolution_clock::now(); - // for (int i = 0; i < runs; ++i) CUDALERP(d_img_tex, oldw, oldh, d_out, neww, newh); - // auto end = high_resolution_clock::now(); - // auto sum = (end - start) / runs; - - // auto h_out = new uint8_t[neww * newh]; - // cudaMemcpy(h_out, d_out, total, cudaMemcpyDeviceToHost); - } +#endif } -- Gitblit v1.8.0