#include "PL_H264Encoder.h" #include "MaterialBuffer.h" #include "logger.h" extern "C" { #include #include #include #include #include } #include PL_H264Encoder_Config::PL_H264Encoder_Config() : inBufferSize(2*1024*1024), // 2MByte resetPTS(false), bytesBufferImageWidth(0), bytesBufferImageHeight(0), avc_bit_rate(1*1024*1024*8), //1Mbit avc_fps(25), avc_gop(25), avc_max_b_frames(0), avc_profile(FF_PROFILE_H264_BASELINE), av_opt_preset("superfast"), av_opt_tune(""), avc_profile_str("") { // av_opt_tune: zerolatency } struct H264Encoder_Internal { uint8_t* buffer; size_t buffSize; bool payError; bool ffmpegInited; size_t frameCount; MB_Frame lastFrame; PL_H264Encoder_Config config; AVCodecContext* pAVCodecContext; AVFrame* pAVFrame;//#todo delete AVFormatContext* pAVFormatContext; H264Encoder_Internal() : buffer(nullptr), buffSize(0), payError(true), ffmpegInited(false), frameCount(0), lastFrame(), config(), pAVCodecContext(nullptr), pAVFrame(nullptr), pAVFormatContext(nullptr) { } ~H264Encoder_Internal() { delete[] buffer; buffer = nullptr; } void reset() { buffSize = 0; payError = true; ffmpegInited = false; frameCount = 0; MB_Frame _lastFrame; lastFrame = _lastFrame; PL_H264Encoder_Config _config; config = _config; pAVCodecContext = nullptr; pAVFrame = nullptr; pAVFormatContext = nullptr; if (buffer != nullptr) { delete[] buffer; buffer = nullptr; } } }; PipeLineElem* create_PL_H264Encoder() { return new PL_H264Encoder; } PL_H264Encoder::PL_H264Encoder() : internal(new H264Encoder_Internal) { } PL_H264Encoder::~PL_H264Encoder() { delete (H264Encoder_Internal*)internal; internal= nullptr; } bool PL_H264Encoder::init(void* args) { H264Encoder_Internal* in = (H264Encoder_Internal*)internal; in->reset(); if (args != nullptr) { PL_H264Encoder_Config* config = (PL_H264Encoder_Config*)args; in->config = *config; } in->buffer = new uint8_t[in->config.inBufferSize]; return true; } void PL_H264Encoder::finit() { H264Encoder_Internal* in = (H264Encoder_Internal*)internal; } bool initH264EncoderEnv(H264Encoder_Internal* in) { av_register_all(); // find the video encoder AVCodec* avCodec = avcodec_find_encoder(AV_CODEC_ID_H264); if (!avCodec) { LOG_ERROR << "codec not found!" << std::endl; return false; } in->pAVCodecContext = avcodec_alloc_context3(avCodec); in->pAVCodecContext->bit_rate = in->config.avc_bit_rate; in->pAVCodecContext->width = in->config.bytesBufferImageWidth; in->pAVCodecContext->height = in->config.bytesBufferImageHeight; in->pAVCodecContext->time_base.num = 1; in->pAVCodecContext->time_base.den = in->config.avc_fps; in->pAVCodecContext->gop_size = in->config.avc_gop; in->pAVCodecContext->max_b_frames = in->config.avc_max_b_frames; in->pAVCodecContext->pix_fmt = AV_PIX_FMT_YUV420P; if (!in->config.av_opt_preset.empty()) av_opt_set(in->pAVCodecContext->priv_data, "preset", in->config.av_opt_preset.c_str(), 0); if (!in->config.av_opt_tune.empty()) av_opt_set(in->pAVCodecContext->priv_data, "tune", in->config.av_opt_tune.c_str(), 0); if (!in->config.avc_profile_str.empty()) av_opt_set(in->pAVCodecContext->priv_data, "profile", in->config.avc_profile_str.c_str(), 0); else in->pAVCodecContext->profile = in->config.avc_profile; if(avcodec_open2(in->pAVCodecContext, avCodec, NULL) >= 0) { in->pAVFrame = av_frame_alloc(); // Allocate video frame in->pAVFrame->format = in->pAVCodecContext->pix_fmt; in->pAVFrame->width = in->pAVCodecContext->width; in->pAVFrame->height = in->pAVCodecContext->height; int ret = av_image_alloc(in->pAVFrame->data, in->pAVFrame->linesize, in->pAVCodecContext->width, in->pAVCodecContext->height, in->pAVCodecContext->pix_fmt, 16); if (ret < 0) { LOG_ERROR << "av_image_alloc error" << std::endl; return false; } } else { LOG_ERROR << "avcodec_open2 error" << std::endl; return false; } return true; } void copyAVFrame(AVFrame* dest, AVFrame* src) { //#test //#define SUBSAMPLE(v, a) ((((v) + (a) - 1)) / (a)) //int src_width = src->width; //int src_height = src->height; //int dst_width = dest->width; //int dst_height = dest->height; //printf("I420Scale sw=%d, sh=%d, dw=%d, dh=%d\n", src_width,src_height,dst_width, dst_height); // //libyuv::I420Scale(src->data[0], src_width, // src->data[1], SUBSAMPLE(src_width, 2), // src->data[2], SUBSAMPLE(src_width, 2), // src_width, src_height, // dest->data[0], dst_width, // dest->data[1], SUBSAMPLE(dst_width, 2), // dest->data[2], SUBSAMPLE(dst_width, 2), // dst_width, dst_height, // libyuv::kFilterNone ); //static size_t f=0; //char fname[50]; //sprintf(fname, "%u.yuv420", ++f); //FILE * pFile = fopen (fname,"wb"); //fwrite (dest->data[0] , sizeof(char), dst_width * dst_height, pFile); //fwrite (dest->data[1] , sizeof(char), dst_width * dst_height / 4, pFile); //fwrite (dest->data[2] , sizeof(char), dst_width * dst_height / 4, pFile); //fclose(pFile); dest->data[0] = src->data[0]; dest->data[1] = src->data[1]; dest->data[2] = src->data[2]; //int height = dest->height; //int width = dest->width; //memcpy(dest->data[0], src->data[0], height * width); // Y //memcpy(dest->data[1], src->data[1], height * width / 4); // U //memcpy(dest->data[2], src->data[2], height * width / 4); // V } bool encodeH264(H264Encoder_Internal* in, AVFrame* pAVFrame, timeval pts) { in->buffSize = 0; copyAVFrame(in->pAVFrame, pAVFrame); //in->pAVFrame->pts = (1.0 / 25) * 90000 * in->frameCount; //in->pAVFrame->pts = time(nullptr); //in->pAVFrame->pts = (pts.tv_sec * 1000 * 1000 + pts.tv_usec) / 90000 + in->frameCount; //in->pAVFrame->pts = (pts.tv_sec * 1000 * 1000 + pts.tv_usec) / 90000 + ((1.0 / 25) * 90000 * in->frameCount); in->pAVFrame->pts = (pts.tv_sec * 1000 * 1000 + pts.tv_usec) / 90000; //in->pAVFrame->pts = pAVFrame->pkt_pts; //in->pAVFrame->pts = (1.0 / 25) * 90000 * in->frameCount; AVPacket pAVPacket = {0}; av_init_packet(&pAVPacket); // encode the image int gotPacket = 0; int ret = avcodec_encode_video2(in->pAVCodecContext, &pAVPacket, in->pAVFrame, &gotPacket); if (ret < 0) { LOG_WARN << "avcodec_encode_video2 (1) error=" << ret << std::endl; return false; } if (gotPacket > 0) { in->frameCount++; LOGP(DEBUG, "Succeed to encode (1) frame=%d, size=%d", in->frameCount, pAVPacket.size); memcpy(in->buffer, pAVPacket.data, pAVPacket.size);//#todo check inBufferSize in->buffSize = pAVPacket.size; av_free_packet(&pAVPacket); } //#todo finit //Flush Encoder, when stop encoder //while (gotPacket > 0) //{ // ret = avcodec_encode_video2(in->pAVCodecContext, &pAVPacket, NULL, &gotPacket); // if (ret < 0) // { // printf("avcodec_encode_video2 (2) error=%d\n", ret); // return false; // } // if (gotPacket > 0) // { // printf("Succeed to encode (2) frame=%d, size=%d\n", in->frameCount, pAVPacket.size); // memcpy(in->buffer + in->buffSize, pAVPacket.data, pAVPacket.size); // in->buffSize += pAVPacket.size; // av_free_packet(&pAVPacket); // } //} //#test //if (in->buffSize > 0) //{ // static FILE * pFile = fopen("out.h264","wba+"); // fwrite (in->buffer , sizeof(char), in->buffSize, pFile); // fflush(pFile); //} in->payError = (in->buffSize == 0); return !(in->payError); } bool encodeH264(H264Encoder_Internal* in, uint8_t* buffer, timeval pts) { uint16_t width = in->config.bytesBufferImageWidth; uint16_t height = in->config.bytesBufferImageHeight; AVFrame avFrame; avFrame.width = width; avFrame.height = height; avFrame.data[0] = buffer; avFrame.data[1] = buffer + width*height; avFrame.data[2] = buffer + width*height + width*height/4; return encodeH264(in, &avFrame, pts); } bool PL_H264Encoder::pay(const PipeMaterial& pm) { H264Encoder_Internal* in = (H264Encoder_Internal*)internal; in->payError = true; if (!in->ffmpegInited) { MB_Frame* frame = (MB_Frame*)pm.buffer; if (frame != nullptr && frame->buffer != nullptr && (in->config.bytesBufferImageWidth == 0 || in->config.bytesBufferImageHeight == 0)) { if (frame->type == MB_Frame::MBFT_PTR_AVFRAME) { AVFrame* pAVFrame = (AVFrame*)frame->buffer; if (pAVFrame != nullptr) { in->config.bytesBufferImageWidth = pAVFrame->width; in->config.bytesBufferImageHeight = pAVFrame->height; LOGP(NOTICE, "Set codec size from AVFrame width=%d, height=%d", in->config.bytesBufferImageWidth, in->config.bytesBufferImageHeight); } } else if (frame->type == MB_Frame::MBFT_YUV420) { in->config.bytesBufferImageWidth = frame->width; in->config.bytesBufferImageHeight = frame->height; LOGP(NOTICE, "Set codec size from frame width=%d, height=%d", in->config.bytesBufferImageWidth, in->config.bytesBufferImageHeight); } } bool ret = initH264EncoderEnv(in); if (!ret) { LOG_ERROR << "initH264EncoderEnv error" << std::endl; return false; } else in->ffmpegInited = true; } if (pm.type != PipeMaterial::PMT_FRAME) { LOG_ERROR << "Only support PMT_FRAME" << std::endl; return false; } if (pm.buffer == nullptr) return false; MB_Frame* frame = (MB_Frame*)pm.buffer; bool ret; if (frame->type == MB_Frame::MBFT_PTR_AVFRAME) ret = encodeH264(in, (AVFrame*)(frame->buffer), frame->pts); else if (frame->type == MB_Frame::MBFT_YUV420) ret = encodeH264(in, (uint8_t*)(frame->buffer), frame->pts); else { LOG_ERROR << "Only support MBFT_PTR_AVFRAME / MBFT_YUV420" << std::endl; in->payError = true; return false; } in->payError = !ret; if (ret) { in->lastFrame.type = MB_Frame::MBFT_H264_NALU; in->lastFrame.buffer = in->buffer; in->lastFrame.buffSize = in->buffSize; in->lastFrame.width = frame->width; in->lastFrame.height = frame->height; if (in->config.resetPTS) gettimeofday(&(in->lastFrame.pts),NULL); else in->lastFrame.pts = frame->pts; } return ret; } bool PL_H264Encoder::gain(PipeMaterial& pm) { H264Encoder_Internal* in = (H264Encoder_Internal*)internal; if (!in->payError) { pm.type = PipeMaterial::PMT_FRAME; pm.buffer = &(in->lastFrame); pm.buffSize = 0; } pm.former = this; return !in->payError; }