博客
关于我
强烈建议你试试无所不能的chatGPT,快点击我
x265-1.7版本-encoder/frameencoder.cpp注释
阅读量:2190 次
发布时间:2019-05-02

本文共 48138 字,大约阅读时间需要 160 分钟。

注:问号以及未注释部分 会在x265-1.8版本内更新

/***************************************************************************** * Copyright (C) 2013 x265 project * * Authors: Chung Shin Yee 
* Min Chen
* Steve Borho
* * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA. * * This program is also available under a commercial proprietary license. * For more information, contact us at license @ x265.com. *****************************************************************************/#include "common.h"#include "frame.h"#include "framedata.h"#include "wavefront.h"#include "param.h"#include "encoder.h"#include "frameencoder.h"#include "common.h"#include "slicetype.h"#include "nal.h"namespace x265 {void weightAnalyse(Slice& slice, Frame& frame, x265_param& param);FrameEncoder::FrameEncoder(){ m_prevOutputTime = x265_mdate(); m_isFrameEncoder = true; m_threadActive = true; m_slicetypeWaitTime = 0; m_activeWorkerCount = 0; m_completionCount = 0; m_bAllRowsStop = false; m_vbvResetTriggerRow = -1; m_outStreams = NULL; m_substreamSizes = NULL; m_nr = NULL; m_tld = NULL; m_rows = NULL; m_top = NULL; m_param = NULL; m_frame = NULL; m_cuGeoms = NULL; m_ctuGeomMap = NULL; m_localTldIdx = 0; memset(&m_frameStats, 0, sizeof(m_frameStats)); memset(&m_rce, 0, sizeof(RateControlEntry));}void FrameEncoder::destroy(){ if (m_pool) { if (!m_jpId) { int numTLD = m_pool->m_numWorkers; if (!m_param->bEnableWavefront) numTLD += m_pool->m_numProviders; for (int i = 0; i < numTLD; i++) m_tld[i].destroy(); delete [] m_tld; } } else { m_tld->destroy(); delete m_tld; } delete[] m_rows; delete[] m_outStreams; X265_FREE(m_cuGeoms); X265_FREE(m_ctuGeomMap); X265_FREE(m_substreamSizes); X265_FREE(m_nr); m_frameFilter.destroy(); if (m_param->bEmitHRDSEI || !!m_param->interlaceMode) { delete m_rce.picTimingSEI; delete m_rce.hrdTiming; }}bool FrameEncoder::init(Encoder *top, int numRows, int numCols){ m_top = top; //获取Encoder类 m_param = top->m_param; m_numRows = numRows; m_numCols = numCols; m_filterRowDelay = (m_param->bEnableSAO && m_param->bSaoNonDeblocked) ? 2 : (m_param->bEnableSAO || m_param->bEnableLoopFilter ? 1 : 0); m_filterRowDelayCus = m_filterRowDelay * numCols; m_rows = new CTURow[m_numRows]; bool ok = !!m_numRows; /* determine full motion search range */ int range = m_param->searchRange; /* fpel search */ range += !!(m_param->searchMethod < 2); /* diamond/hex range check lag */ range += NTAPS_LUMA / 2; /* subpel filter half-length */ range += 2 + MotionEstimate::hpelIterationCount(m_param->subpelRefine) / 2; /* subpel refine steps */ m_refLagRows = 1 + ((range + g_maxCUSize - 1) / g_maxCUSize); // NOTE: 2 times of numRows because both Encoder and Filter in same queue if (!WaveFront::init(m_numRows * 2)) { x265_log(m_param, X265_LOG_ERROR, "unable to initialize wavefront queue\n"); m_pool = NULL; } m_frameFilter.init(top, this, numRows); // initialize HRD parameters of SPS if (m_param->bEmitHRDSEI || !!m_param->interlaceMode) { m_rce.picTimingSEI = new SEIPictureTiming; m_rce.hrdTiming = new HRDTiming; ok &= m_rce.picTimingSEI && m_rce.hrdTiming; } if (m_param->noiseReductionIntra || m_param->noiseReductionInter) m_nr = X265_MALLOC(NoiseReduction, 1); if (m_nr) memset(m_nr, 0, sizeof(NoiseReduction)); else m_param->noiseReductionIntra = m_param->noiseReductionInter = 0; return ok;}/* Generate a complete list of unique geom sets for the current picture dimensions *//** 函数功能 : 计算CU所以情况的几何信息/* 调用范围 : 只在FrameEncoder::startCompressFrame函数中被调用* 返回值 : 成功返回ture 失败返回 false**/bool FrameEncoder::initializeGeoms(){ /* Geoms only vary between CTUs in the presence of picture edges */ int maxCUSize = m_param->maxCUSize;//最大CTU int minCUSize = m_param->minCUSize;;//最小CTU int heightRem = m_param->sourceHeight & (maxCUSize - 1);//高度不够CTU的余数 int widthRem = m_param->sourceWidth & (maxCUSize - 1);//宽度不够CTU的余数 int allocGeoms = 1; // body //存储的个数:分别为:CTU中全部有像素值 CTU右边不够像素值 CTU下边不够像素值 CTU右边和下边不够像素值 if (heightRem && widthRem) allocGeoms = 4; // body, right, bottom, corner else if (heightRem || widthRem) allocGeoms = 2; // body, right or bottom m_ctuGeomMap = X265_MALLOC(uint32_t, m_numRows * m_numCols); m_cuGeoms = X265_MALLOC(CUGeom, allocGeoms * CUGeom::MAX_GEOMS); if (!m_cuGeoms || !m_ctuGeomMap) return false; // body CUData::calcCTUGeoms(maxCUSize, maxCUSize, maxCUSize, minCUSize, m_cuGeoms);//申请 CTU中全部有像素值 部分 memset(m_ctuGeomMap, 0, sizeof(uint32_t) * m_numRows * m_numCols); if (allocGeoms == 1) return true; int countGeoms = 1; if (widthRem) { // right CUData::calcCTUGeoms(widthRem, maxCUSize, maxCUSize, minCUSize, m_cuGeoms + countGeoms * CUGeom::MAX_GEOMS); //申请 CTU右边不够像素值 部分 for (uint32_t i = 0; i < m_numRows; i++) { uint32_t ctuAddr = m_numCols * (i + 1) - 1; m_ctuGeomMap[ctuAddr] = countGeoms * CUGeom::MAX_GEOMS; } countGeoms++; } if (heightRem) { // bottom CUData::calcCTUGeoms(maxCUSize, heightRem, maxCUSize, minCUSize, m_cuGeoms + countGeoms * CUGeom::MAX_GEOMS);//申请 CTU下边不够像素值 部分 for (uint32_t i = 0; i < m_numCols; i++) { uint32_t ctuAddr = m_numCols * (m_numRows - 1) + i; m_ctuGeomMap[ctuAddr] = countGeoms * CUGeom::MAX_GEOMS; } countGeoms++; if (widthRem) { // corner CUData::calcCTUGeoms(widthRem, heightRem, maxCUSize, minCUSize, m_cuGeoms + countGeoms * CUGeom::MAX_GEOMS);//申请 CTU右边和下边不够像素值 部分 uint32_t ctuAddr = m_numCols * m_numRows - 1; m_ctuGeomMap[ctuAddr] = countGeoms * CUGeom::MAX_GEOMS; countGeoms++; } X265_CHECK(countGeoms == allocGeoms, "geometry match check failure\n"); } return true;}/** 函数功能 : 触发compressframe()进行编码/* 调用范围 : 只在Encoder::encode函数中被调用* \参数 curFrame : 待编码帧* 返回值 : 成功返回true 异常返回false**/bool FrameEncoder::startCompressFrame(Frame* curFrame){ m_slicetypeWaitTime = x265_mdate() - m_prevOutputTime;//计算从上一帧编码完毕到开始编码新一帧的等待时间 m_frame = curFrame;//设置当前处理的帧 m_param = curFrame->m_param;//获取配置参数 m_sliceType = curFrame->m_lowres.sliceType;//获取当前的帧类型 curFrame->m_encData->m_frameEncoderID = m_jpId;//获取当前的job id curFrame->m_encData->m_jobProvider = this;//让当前帧获取当前所在线程的位置 curFrame->m_encData->m_slice->m_mref = m_mref;//获取参考帧信息 if (!m_cuGeoms)//编码器只执行framethread个数,申请所有情况CU的几何信息 { if (!initializeGeoms())//计算CU所以情况的几何信息 return false; } m_enable.trigger();//当前帧的帧类型信息 已经确定完毕 触发准备编码 return true;}/** 函数功能 : ??触发compressframe()进行编码/* 调用范围 : 线程循环触发,encoder->create() frame start开始 (frameThread配置多少,起几个线程)* 返回值 : null**/void FrameEncoder::threadMain(){ THREAD_NAME("Frame", m_jpId); //本函数会被执行配置的framethread次数,因为起了framethread个线程进行compressframe if (m_pool) { m_pool->setCurrentThreadAffinity();//设置线程间能够在不同的核运行,而不会同时占用同一个核 /* the first FE on each NUMA node is responsible for allocating thread * local data for all worker threads in that pool. If WPP is disabled, then * each FE also needs a TLD instance */ if (!m_jpId) //只在m_jpid = 0 时才会进入,也就说在framethread个线程中只有第一个才会进入 (因为这次初始化为把所有并行的analysis都初始化) { int numTLD = m_pool->m_numWorkers; //获取当前机器核数 单机4核测试是4 if (!m_param->bEnableWavefront) //如果关闭WPP numTLD += m_pool->m_numProviders; m_tld = new ThreadLocalData[numTLD];//申请并行空间 for (int i = 0; i < numTLD; i++) { m_tld[i].analysis.initSearch(*m_param, m_top->m_scalingList);//??? m_tld[i].analysis.create(m_tld);//??? } //将所有的Encoder的m_tld都指向第一个m_tld for (int i = 0; i < m_pool->m_numProviders; i++) //遍历所有线程任务 { if (m_pool->m_jpTable[i]->m_isFrameEncoder) //如果当前 是frameEcnoder 不是lookachead /* ugh; over-allocation and other issues here */ { FrameEncoder *peer = dynamic_cast
(m_pool->m_jpTable[i]);//强制转换类型 peer->m_tld = m_tld;//公用同一ThreadLocalData buffer } } } if (m_param->bEnableWavefront) m_localTldIdx = -1; //??? cause exception if used else m_localTldIdx = m_pool->m_numWorkers + m_jpId;//????? } else//不使用并行,直接单线程初始化 { m_tld = new ThreadLocalData; m_tld->analysis.initSearch(*m_param, m_top->m_scalingList); m_tld->analysis.create(NULL); m_localTldIdx = 0; } m_done.trigger(); /* signal that thread is initialized */ m_enable.wait(); /* Encoder::encode() triggers this event */ //m_done在encoder.create中先进行wait 每完成一帧即compressFrame()之后才会触发 //以上两个是一个PV原语,m_enable表示准备开始编码,在startCompressFrame函数中触发在compressFrame之后wait //m_done表示编码完成,在getEncodedPicture之前wait,在compressFrame之后触发 while (m_threadActive)//循环执行任务,知道encoder->stop 中设置结束 { compressFrame();//编码视频 m_done.trigger();//编码完毕触发完成 /* FrameEncoder::getEncodedPicture() blocks for this event */ m_enable.wait();//等待下一帧是否准备 }}/** 函数功能 : 分析加权信息/* 调用范围 : 只在WorkerThread::threadMain()函数中被调用* \返回 : null * */void FrameEncoder::WeightAnalysis::processTasks(int /* workerThreadId */){ Frame* frame = master.m_frame;//获取当前分析帧 weightAnalyse(*frame->m_encData->m_slice, *frame, *master.m_param);//分析加权信息}/** 函数功能 : ??触发compressframe()进行编码/* 调用范围 : 只在FrameEncoder::threadMain()函数中被调用* 返回值 : null**/void FrameEncoder::compressFrame(){ ProfileScopeEvent(frameThread); //功能:???? // 1. 初始化数据 // 2. NAL?????? // 3. 参考帧加权分析 // 4. 配置参考帧信息 m_startCompressTime = x265_mdate();//获取开始编码的时间点 m_totalActiveWorkerCount = 0;//??? m_activeWorkerCountSamples = 0;//??? m_totalWorkerElapsedTime = 0;//??? m_totalNoWorkerTime = 0;//??? m_countRowBlocks = 0;//??? m_allRowsAvailableTime = 0;//??? m_stallStartTime = 0;//??? m_completionCount = 0;//??? m_bAllRowsStop = false;//??? m_vbvResetTriggerRow = -1;//??? m_SSDY = m_SSDU = m_SSDV = 0;//??? m_ssim = 0;//??? m_ssimCnt = 0;//??? memset(&m_frameStats, 0, sizeof(m_frameStats));//??? /* Emit access unit delimiter unless this is the first frame and the user is * not repeating headers (since AUD is supposed to be the first NAL in the access * unit) */ Slice* slice = m_frame->m_encData->m_slice;//获取当前slice if (m_param->bEnableAccessUnitDelimiters && (m_frame->m_poc || m_param->bRepeatHeaders))//???? { m_bs.resetBits(); m_entropyCoder.setBitstream(&m_bs); m_entropyCoder.codeAUD(*slice); m_bs.writeByteAlignment(); m_nalList.serialize(NAL_UNIT_ACCESS_UNIT_DELIMITER, m_bs); } if (m_frame->m_lowres.bKeyframe && m_param->bRepeatHeaders)//???? m_top->getStreamHeaders(m_nalList, m_entropyCoder, m_bs); // Weighted Prediction parameters estimation. bool bUseWeightP = slice->m_sliceType == P_SLICE && slice->m_pps->bUseWeightPred; //当前是否应用P帧加权预测 bool bUseWeightB = slice->m_sliceType == B_SLICE && slice->m_pps->bUseWeightedBiPred;//当前是否应用B帧加权预测 if (bUseWeightP || bUseWeightB) {#if DETAILED_CU_STATS m_cuStats.countWeightAnalyze++; ScopedElapsedTime time(m_cuStats.weightAnalyzeTime);#endif WeightAnalysis wa(*this);//用于多线程 加权分析 if (m_pool && wa.tryBondPeers(*this, 1))//从当前job中拥有核并且sleep状态的核可以触发多线程,如果没有可用核则在当前线程中完成进入else /* use an idle worker for weight analysis */ wa.waitForExit();//一直等待到任务全部完成,这里等待的是核释放,内核释放了任务也就完成了 else weightAnalyse(*slice, *m_frame, *m_param);//分析加权信息(每个list的第一帧分析加权与否,其它不加权) } else slice->disableWeights();//关闭当前帧的加权预测 // Generate motion references //配置参考帧信息 int numPredDir = slice->isInterP() ? 1 : slice->isInterB() ? 2 : 0;//获取当前有几个list for (int l = 0; l < numPredDir; l++) //遍历两个list { for (int ref = 0; ref < slice->m_numRefIdx[l]; ref++)//遍历当前list的所有参考帧 { WeightParam *w = NULL; if ((bUseWeightP || bUseWeightB) && slice->m_weightPredTable[l][ref][0].bPresentFlag)//如果当前应用加权预测,并且前面分析加权优 w = slice->m_weightPredTable[l][ref];//获取加权参数 m_mref[l][ref].init(slice->m_refPicList[l][ref]->m_reconPic, w, *m_param);//获取参考帧信息,申请加权帧内存 } } /* Get the QP for this frame from rate control. This call may block until * frames ahead of it in encode order have called rateControlEnd() */ int qp = m_top->m_rateControl->rateControlStart(m_frame, &m_rce, m_top);//??? m_rce.newQp = qp; /* Clip slice QP to 0-51 spec range before encoding */ slice->m_sliceQp = x265_clip3(-QP_BD_OFFSET, QP_MAX_SPEC, qp); m_initSliceContext.resetEntropy(*slice); m_frameFilter.start(m_frame, m_initSliceContext, qp); /* ensure all rows are blocked prior to initializing row CTU counters */ WaveFront::clearEnabledRowMask(); /* reset entropy coders */ m_entropyCoder.load(m_initSliceContext); for (uint32_t i = 0; i < m_numRows; i++) m_rows[i].init(m_initSliceContext); uint32_t numSubstreams = m_param->bEnableWavefront ? slice->m_sps->numCuInHeight : 1; if (!m_outStreams) { m_outStreams = new Bitstream[numSubstreams]; m_substreamSizes = X265_MALLOC(uint32_t, numSubstreams); if (!m_param->bEnableSAO) for (uint32_t i = 0; i < numSubstreams; i++) m_rows[i].rowGoOnCoder.setBitstream(&m_outStreams[i]); } else for (uint32_t i = 0; i < numSubstreams; i++) m_outStreams[i].resetBits(); int prevBPSEI = m_rce.encodeOrder ? m_top->m_lastBPSEI : 0; if (m_frame->m_lowres.bKeyframe) { if (m_param->bEmitHRDSEI) { SEIBufferingPeriod* bpSei = &m_top->m_rateControl->m_bufPeriodSEI; // since the temporal layer HRD is not ready, we assumed it is fixed bpSei->m_auCpbRemovalDelayDelta = 1; bpSei->m_cpbDelayOffset = 0; bpSei->m_dpbDelayOffset = 0; // hrdFullness() calculates the initial CPB removal delay and offset m_top->m_rateControl->hrdFullness(bpSei); m_bs.resetBits(); bpSei->write(m_bs, *slice->m_sps); m_bs.writeByteAlignment(); m_nalList.serialize(NAL_UNIT_PREFIX_SEI, m_bs); m_top->m_lastBPSEI = m_rce.encodeOrder; } // The recovery point SEI message assists a decoder in determining when the decoding // process will produce acceptable pictures for display after the decoder initiates // random access. The m_recoveryPocCnt is in units of POC(picture order count) which // means pictures encoded after the CRA but precede it in display order(leading) are // implicitly discarded after a random access seek regardless of the value of // m_recoveryPocCnt. Our encoder does not use references prior to the most recent CRA, // so all pictures following the CRA in POC order are guaranteed to be displayable, // so m_recoveryPocCnt is always 0. SEIRecoveryPoint sei_recovery_point; sei_recovery_point.m_recoveryPocCnt = 0; sei_recovery_point.m_exactMatchingFlag = true; sei_recovery_point.m_brokenLinkFlag = false; m_bs.resetBits(); sei_recovery_point.write(m_bs, *slice->m_sps); m_bs.writeByteAlignment(); m_nalList.serialize(NAL_UNIT_PREFIX_SEI, m_bs); } if (m_param->bEmitHRDSEI || !!m_param->interlaceMode) { SEIPictureTiming *sei = m_rce.picTimingSEI; const VUI *vui = &slice->m_sps->vuiParameters; const HRDInfo *hrd = &vui->hrdParameters; int poc = slice->m_poc; if (vui->frameFieldInfoPresentFlag) { if (m_param->interlaceMode == 2) sei->m_picStruct = (poc & 1) ? 1 /* top */ : 2 /* bottom */; else if (m_param->interlaceMode == 1) sei->m_picStruct = (poc & 1) ? 2 /* bottom */ : 1 /* top */; else sei->m_picStruct = 0; sei->m_sourceScanType = 0; sei->m_duplicateFlag = false; } if (vui->hrdParametersPresentFlag) { // The m_aucpbremoval delay specifies how many clock ticks the // access unit associated with the picture timing SEI message has to // wait after removal of the access unit with the most recent // buffering period SEI message sei->m_auCpbRemovalDelay = X265_MIN(X265_MAX(1, m_rce.encodeOrder - prevBPSEI), (1 << hrd->cpbRemovalDelayLength)); sei->m_picDpbOutputDelay = slice->m_sps->numReorderPics + poc - m_rce.encodeOrder; } m_bs.resetBits(); sei->write(m_bs, *slice->m_sps); m_bs.writeByteAlignment(); m_nalList.serialize(NAL_UNIT_PREFIX_SEI, m_bs); } /* Analyze CTU rows, most of the hard work is done here. Frame is * compressed in a wave-front pattern if WPP is enabled. Row based loop * filters runs behind the CTU compression and reconstruction */ m_rows[0].active = true; if (m_param->bEnableWavefront) { for (uint32_t row = 0; row < m_numRows; row++) { // block until all reference frames have reconstructed the rows we need for (int l = 0; l < numPredDir; l++) { for (int ref = 0; ref < slice->m_numRefIdx[l]; ref++) { Frame *refpic = slice->m_refPicList[l][ref]; uint32_t reconRowCount = refpic->m_reconRowCount.get(); while ((reconRowCount != m_numRows) && (reconRowCount < row + m_refLagRows)) reconRowCount = refpic->m_reconRowCount.waitForChange(reconRowCount); if ((bUseWeightP || bUseWeightB) && m_mref[l][ref].isWeighted) m_mref[l][ref].applyWeight(row + m_refLagRows, m_numRows); } } enableRowEncoder(row); /* clear external dependency for this row */ if (!row) { m_row0WaitTime = x265_mdate(); enqueueRowEncoder(0); /* clear internal dependency, start wavefront */ } tryWakeOne(); } m_allRowsAvailableTime = x265_mdate(); tryWakeOne(); /* ensure one thread is active or help-wanted flag is set prior to blocking */ static const int block_ms = 250; while (m_completionEvent.timedWait(block_ms)) tryWakeOne(); } else { for (uint32_t i = 0; i < m_numRows + m_filterRowDelay; i++) { // compress if (i < m_numRows) { // block until all reference frames have reconstructed the rows we need for (int l = 0; l < numPredDir; l++) { int list = l; for (int ref = 0; ref < slice->m_numRefIdx[list]; ref++) { Frame *refpic = slice->m_refPicList[list][ref]; uint32_t reconRowCount = refpic->m_reconRowCount.get(); while ((reconRowCount != m_numRows) && (reconRowCount < i + m_refLagRows)) reconRowCount = refpic->m_reconRowCount.waitForChange(reconRowCount); if ((bUseWeightP || bUseWeightB) && m_mref[l][ref].isWeighted) m_mref[list][ref].applyWeight(i + m_refLagRows, m_numRows); } } if (!i) m_row0WaitTime = x265_mdate(); else if (i == m_numRows - 1) m_allRowsAvailableTime = x265_mdate(); processRowEncoder(i, m_tld[m_localTldIdx]); } // filter if (i >= m_filterRowDelay) m_frameFilter.processRow(i - m_filterRowDelay); } } if (m_param->rc.bStatWrite) { int totalI = 0, totalP = 0, totalSkip = 0; // accumulate intra,inter,skip cu count per frame for 2 pass for (uint32_t i = 0; i < m_numRows; i++) { m_frameStats.mvBits += m_rows[i].rowStats.mvBits; m_frameStats.coeffBits += m_rows[i].rowStats.coeffBits; m_frameStats.miscBits += m_rows[i].rowStats.miscBits; totalI += m_rows[i].rowStats.iCuCnt; totalP += m_rows[i].rowStats.pCuCnt; totalSkip += m_rows[i].rowStats.skipCuCnt; } int totalCuCount = totalI + totalP + totalSkip; m_frameStats.percentIntra = (double)totalI / totalCuCount; m_frameStats.percentInter = (double)totalP / totalCuCount; m_frameStats.percentSkip = (double)totalSkip / totalCuCount; } m_bs.resetBits(); m_entropyCoder.load(m_initSliceContext); m_entropyCoder.setBitstream(&m_bs); m_entropyCoder.codeSliceHeader(*slice, *m_frame->m_encData); // finish encode of each CTU row, only required when SAO is enabled if (m_param->bEnableSAO) encodeSlice(); // serialize each row, record final lengths in slice header uint32_t maxStreamSize = m_nalList.serializeSubstreams(m_substreamSizes, numSubstreams, m_outStreams); // complete the slice header by writing WPP row-starts m_entropyCoder.setBitstream(&m_bs); if (slice->m_pps->bEntropyCodingSyncEnabled) m_entropyCoder.codeSliceHeaderWPPEntryPoints(*slice, m_substreamSizes, maxStreamSize); m_bs.writeByteAlignment(); m_nalList.serialize(slice->m_nalUnitType, m_bs); if (m_param->decodedPictureHashSEI) { if (m_param->decodedPictureHashSEI == 1) { m_seiReconPictureDigest.m_method = SEIDecodedPictureHash::MD5; for (int i = 0; i < 3; i++) MD5Final(&m_state[i], m_seiReconPictureDigest.m_digest[i]); } else if (m_param->decodedPictureHashSEI == 2) { m_seiReconPictureDigest.m_method = SEIDecodedPictureHash::CRC; for (int i = 0; i < 3; i++) crcFinish(m_crc[i], m_seiReconPictureDigest.m_digest[i]); } else if (m_param->decodedPictureHashSEI == 3) { m_seiReconPictureDigest.m_method = SEIDecodedPictureHash::CHECKSUM; for (int i = 0; i < 3; i++) checksumFinish(m_checksum[i], m_seiReconPictureDigest.m_digest[i]); } m_bs.resetBits(); m_seiReconPictureDigest.write(m_bs, *slice->m_sps); m_bs.writeByteAlignment(); m_nalList.serialize(NAL_UNIT_SUFFIX_SEI, m_bs); } uint64_t bytes = 0; for (uint32_t i = 0; i < m_nalList.m_numNal; i++) { int type = m_nalList.m_nal[i].type; // exclude SEI if (type != NAL_UNIT_PREFIX_SEI && type != NAL_UNIT_SUFFIX_SEI) { bytes += m_nalList.m_nal[i].sizeBytes; // and exclude start code prefix bytes -= (!i || type == NAL_UNIT_SPS || type == NAL_UNIT_PPS) ? 4 : 3; } } m_accessUnitBits = bytes << 3; m_endCompressTime = x265_mdate(); /* rateControlEnd may also block for earlier frames to call rateControlUpdateStats */ if (m_top->m_rateControl->rateControlEnd(m_frame, m_accessUnitBits, &m_rce, &m_frameStats) < 0) m_top->m_aborted = true; /* Decrement referenced frame reference counts, allow them to be recycled */ for (int l = 0; l < numPredDir; l++) { for (int ref = 0; ref < slice->m_numRefIdx[l]; ref++) { Frame *refpic = slice->m_refPicList[l][ref]; ATOMIC_DEC(&refpic->m_countRefEncoders); } } int numTLD; if (m_pool) numTLD = m_param->bEnableWavefront ? m_pool->m_numWorkers : m_pool->m_numWorkers + m_pool->m_numProviders; else numTLD = 1; if (m_nr) { /* Accumulate NR statistics from all worker threads */ for (int i = 0; i < numTLD; i++) { NoiseReduction* nr = &m_tld[i].analysis.m_quant.m_frameNr[m_jpId]; for (int cat = 0; cat < MAX_NUM_TR_CATEGORIES; cat++) { for (int coeff = 0; coeff < MAX_NUM_TR_COEFFS; coeff++) m_nr->residualSum[cat][coeff] += nr->residualSum[cat][coeff]; m_nr->count[cat] += nr->count[cat]; } } noiseReductionUpdate(); /* Copy updated NR coefficients back to all worker threads */ for (int i = 0; i < numTLD; i++) { NoiseReduction* nr = &m_tld[i].analysis.m_quant.m_frameNr[m_jpId]; memcpy(nr->offsetDenoise, m_nr->offsetDenoise, sizeof(uint16_t) * MAX_NUM_TR_CATEGORIES * MAX_NUM_TR_COEFFS); memset(nr->count, 0, sizeof(uint32_t) * MAX_NUM_TR_CATEGORIES); memset(nr->residualSum, 0, sizeof(uint32_t) * MAX_NUM_TR_CATEGORIES * MAX_NUM_TR_COEFFS); } }#if DETAILED_CU_STATS /* Accumulate CU statistics from each worker thread, we could report * per-frame stats here, but currently we do not. */ for (int i = 0; i < numTLD; i++) m_cuStats.accumulate(m_tld[i].analysis.m_stats[m_jpId]);#endif m_endFrameTime = x265_mdate();}void FrameEncoder::encodeSlice(){ Slice* slice = m_frame->m_encData->m_slice; const uint32_t widthInLCUs = slice->m_sps->numCuInWidth; const uint32_t lastCUAddr = (slice->m_endCUAddr + NUM_4x4_PARTITIONS - 1) / NUM_4x4_PARTITIONS; const uint32_t numSubstreams = m_param->bEnableWavefront ? slice->m_sps->numCuInHeight : 1; SAOParam* saoParam = slice->m_sps->bUseSAO ? m_frame->m_encData->m_saoParam : NULL; for (uint32_t cuAddr = 0; cuAddr < lastCUAddr; cuAddr++) { uint32_t col = cuAddr % widthInLCUs; uint32_t lin = cuAddr / widthInLCUs; uint32_t subStrm = lin % numSubstreams; CUData* ctu = m_frame->m_encData->getPicCTU(cuAddr); m_entropyCoder.setBitstream(&m_outStreams[subStrm]); // Synchronize cabac probabilities with upper-right CTU if it's available and we're at the start of a line. if (m_param->bEnableWavefront && !col && lin) { m_entropyCoder.copyState(m_initSliceContext); m_entropyCoder.loadContexts(m_rows[lin - 1].bufferedEntropy); } if (saoParam) { if (saoParam->bSaoFlag[0] || saoParam->bSaoFlag[1]) { int mergeLeft = col && saoParam->ctuParam[0][cuAddr].mergeMode == SAO_MERGE_LEFT; int mergeUp = lin && saoParam->ctuParam[0][cuAddr].mergeMode == SAO_MERGE_UP; if (col) m_entropyCoder.codeSaoMerge(mergeLeft); if (lin && !mergeLeft) m_entropyCoder.codeSaoMerge(mergeUp); if (!mergeLeft && !mergeUp) { if (saoParam->bSaoFlag[0]) m_entropyCoder.codeSaoOffset(saoParam->ctuParam[0][cuAddr], 0); if (saoParam->bSaoFlag[1]) { m_entropyCoder.codeSaoOffset(saoParam->ctuParam[1][cuAddr], 1); m_entropyCoder.codeSaoOffset(saoParam->ctuParam[2][cuAddr], 2); } } } else { for (int i = 0; i < 3; i++) saoParam->ctuParam[i][cuAddr].reset(); } } // final coding (bitstream generation) for this CU m_entropyCoder.encodeCTU(*ctu, m_cuGeoms[m_ctuGeomMap[cuAddr]]); if (m_param->bEnableWavefront) { if (col == 1) // Store probabilities of second CTU in line into buffer m_rows[lin].bufferedEntropy.loadContexts(m_entropyCoder); if (col == widthInLCUs - 1) m_entropyCoder.finishSlice(); } } if (!m_param->bEnableWavefront) m_entropyCoder.finishSlice();}void FrameEncoder::processRow(int row, int threadId){ int64_t startTime = x265_mdate(); if (ATOMIC_INC(&m_activeWorkerCount) == 1 && m_stallStartTime) m_totalNoWorkerTime += x265_mdate() - m_stallStartTime; const uint32_t realRow = row >> 1; const uint32_t typeNum = row & 1; if (!typeNum) processRowEncoder(realRow, m_tld[threadId]); else { m_frameFilter.processRow(realRow); // NOTE: Active next row if (realRow != m_numRows - 1) enqueueRowFilter(realRow + 1); } if (ATOMIC_DEC(&m_activeWorkerCount) == 0) m_stallStartTime = x265_mdate(); m_totalWorkerElapsedTime += x265_mdate() - startTime; // not thread safe, but good enough}// Called by worker threadsvoid FrameEncoder::processRowEncoder(int intRow, ThreadLocalData& tld){ uint32_t row = (uint32_t)intRow; CTURow& curRow = m_rows[row]; tld.analysis.m_param = m_param; if (m_param->bEnableWavefront) { ScopedLock self(curRow.lock); if (!curRow.active) /* VBV restart is in progress, exit out */ return; if (curRow.busy) { /* On multi-socket Windows servers, we have seen problems with * ATOMIC_CAS which resulted in multiple worker threads processing * the same CU row, which often resulted in bad pointer accesses. We * believe the problem is fixed, but are leaving this check in place * to prevent crashes in case it is not */ x265_log(m_param, X265_LOG_WARNING, "internal error - simultaneous row access detected. Please report HW to x265-devel@videolan.org\n"); return; } curRow.busy = true; } /* When WPP is enabled, every row has its own row coder instance. Otherwise * they share row 0 */ Entropy& rowCoder = m_param->bEnableWavefront ? m_rows[row].rowGoOnCoder : m_rows[0].rowGoOnCoder; FrameData& curEncData = *m_frame->m_encData; Slice *slice = curEncData.m_slice; const uint32_t numCols = m_numCols; const uint32_t lineStartCUAddr = row * numCols; bool bIsVbv = m_param->rc.vbvBufferSize > 0 && m_param->rc.vbvMaxBitrate > 0; /* These store the count of inter, intra and skip cus within quad tree structure of each CTU */ uint32_t qTreeInterCnt[NUM_CU_DEPTH]; uint32_t qTreeIntraCnt[NUM_CU_DEPTH]; uint32_t qTreeSkipCnt[NUM_CU_DEPTH]; for (uint32_t depth = 0; depth <= g_maxCUDepth; depth++) qTreeIntraCnt[depth] = qTreeInterCnt[depth] = qTreeSkipCnt[depth] = 0; while (curRow.completed < numCols) { ProfileScopeEvent(encodeCTU); uint32_t col = curRow.completed; const uint32_t cuAddr = lineStartCUAddr + col; CUData* ctu = curEncData.getPicCTU(cuAddr); ctu->initCTU(*m_frame, cuAddr, slice->m_sliceQp); if (bIsVbv) { if (!row) { curEncData.m_rowStat[row].diagQp = curEncData.m_avgQpRc; curEncData.m_rowStat[row].diagQpScale = x265_qp2qScale(curEncData.m_avgQpRc); } FrameData::RCStatCU& cuStat = curEncData.m_cuStat[cuAddr]; if (row >= col && row && m_vbvResetTriggerRow != intRow) cuStat.baseQp = curEncData.m_cuStat[cuAddr - numCols + 1].baseQp; else cuStat.baseQp = curEncData.m_rowStat[row].diagQp; /* TODO: use defines from slicetype.h for lowres block size */ uint32_t maxBlockCols = (m_frame->m_fencPic->m_picWidth + (16 - 1)) / 16; uint32_t maxBlockRows = (m_frame->m_fencPic->m_picHeight + (16 - 1)) / 16; uint32_t noOfBlocks = g_maxCUSize / 16; uint32_t block_y = (cuAddr / curEncData.m_slice->m_sps->numCuInWidth) * noOfBlocks; uint32_t block_x = (cuAddr * noOfBlocks) - block_y * curEncData.m_slice->m_sps->numCuInWidth; cuStat.vbvCost = 0; cuStat.intraVbvCost = 0; for (uint32_t h = 0; h < noOfBlocks && block_y < maxBlockRows; h++, block_y++) { uint32_t idx = block_x + (block_y * maxBlockCols); for (uint32_t w = 0; w < noOfBlocks && (block_x + w) < maxBlockCols; w++, idx++) { cuStat.vbvCost += m_frame->m_lowres.lowresCostForRc[idx] & LOWRES_COST_MASK; cuStat.intraVbvCost += m_frame->m_lowres.intraCost[idx]; } } } else curEncData.m_cuStat[cuAddr].baseQp = curEncData.m_avgQpRc; if (m_param->bEnableWavefront && !col && row) { // Load SBAC coder context from previous row and initialize row state. rowCoder.copyState(m_initSliceContext); rowCoder.loadContexts(m_rows[row - 1].bufferedEntropy); } // Does all the CU analysis, returns best top level mode decision Mode& best = tld.analysis.compressCTU(*ctu, *m_frame, m_cuGeoms[m_ctuGeomMap[cuAddr]], rowCoder); // take a sample of the current active worker count ATOMIC_ADD(&m_totalActiveWorkerCount, m_activeWorkerCount); ATOMIC_INC(&m_activeWorkerCountSamples); /* advance top-level row coder to include the context of this CTU. * if SAO is disabled, rowCoder writes the final CTU bitstream */ rowCoder.encodeCTU(*ctu, m_cuGeoms[m_ctuGeomMap[cuAddr]]); if (m_param->bEnableWavefront && col == 1) // Save CABAC state for next row curRow.bufferedEntropy.loadContexts(rowCoder); // Completed CU processing curRow.completed++; if (m_param->bLogCuStats || m_param->rc.bStatWrite) curEncData.m_rowStat[row].sumQpAq += collectCTUStatistics(*ctu, qTreeInterCnt, qTreeIntraCnt, qTreeSkipCnt); else if (m_param->rc.aqMode) curEncData.m_rowStat[row].sumQpAq += calcCTUQP(*ctu); // copy no. of intra, inter Cu cnt per row into frame stats for 2 pass if (m_param->rc.bStatWrite) { curRow.rowStats.mvBits += best.mvBits; curRow.rowStats.coeffBits += best.coeffBits; curRow.rowStats.miscBits += best.totalBits - (best.mvBits + best.coeffBits); for (uint32_t depth = 0; depth <= g_maxCUDepth; depth++) { /* 1 << shift == number of 8x8 blocks at current depth */ int shift = 2 * (g_maxCUDepth - depth); curRow.rowStats.iCuCnt += qTreeIntraCnt[depth] << shift; curRow.rowStats.pCuCnt += qTreeInterCnt[depth] << shift; curRow.rowStats.skipCuCnt += qTreeSkipCnt[depth] << shift; // clear the row cu data from thread local object qTreeIntraCnt[depth] = qTreeInterCnt[depth] = qTreeSkipCnt[depth] = 0; } } curEncData.m_cuStat[cuAddr].totalBits = best.totalBits; x265_emms(); if (bIsVbv) { // Update encoded bits, satdCost, baseQP for each CU curEncData.m_rowStat[row].diagSatd += curEncData.m_cuStat[cuAddr].vbvCost; curEncData.m_rowStat[row].diagIntraSatd += curEncData.m_cuStat[cuAddr].intraVbvCost; curEncData.m_rowStat[row].encodedBits += curEncData.m_cuStat[cuAddr].totalBits; curEncData.m_rowStat[row].sumQpRc += curEncData.m_cuStat[cuAddr].baseQp; curEncData.m_rowStat[row].numEncodedCUs = cuAddr; // If current block is at row diagonal checkpoint, call vbv ratecontrol. if (row == col && row) { double qpBase = curEncData.m_cuStat[cuAddr].baseQp; int reEncode = m_top->m_rateControl->rowDiagonalVbvRateControl(m_frame, row, &m_rce, qpBase); qpBase = x265_clip3((double)QP_MIN, (double)QP_MAX_MAX, qpBase); curEncData.m_rowStat[row].diagQp = qpBase; curEncData.m_rowStat[row].diagQpScale = x265_qp2qScale(qpBase); if (reEncode < 0) { x265_log(m_param, X265_LOG_DEBUG, "POC %d row %d - encode restart required for VBV, to %.2f from %.2f\n", m_frame->m_poc, row, qpBase, curEncData.m_cuStat[cuAddr].baseQp); // prevent the WaveFront::findJob() method from providing new jobs m_vbvResetTriggerRow = row; m_bAllRowsStop = true; for (uint32_t r = m_numRows - 1; r >= row; r--) { CTURow& stopRow = m_rows[r]; if (r != row) { /* if row was active (ready to be run) clear active bit and bitmap bit for this row */ stopRow.lock.acquire(); while (stopRow.active) { if (dequeueRow(r * 2)) stopRow.active = false; else { /* we must release the row lock to allow the thread to exit */ stopRow.lock.release(); GIVE_UP_TIME(); stopRow.lock.acquire(); } } stopRow.lock.release(); bool bRowBusy = true; do { stopRow.lock.acquire(); bRowBusy = stopRow.busy; stopRow.lock.release(); if (bRowBusy) { GIVE_UP_TIME(); } } while (bRowBusy); } m_outStreams[r].resetBits(); stopRow.completed = 0; memset(&stopRow.rowStats, 0, sizeof(stopRow.rowStats)); curEncData.m_rowStat[r].numEncodedCUs = 0; curEncData.m_rowStat[r].encodedBits = 0; curEncData.m_rowStat[r].diagSatd = 0; curEncData.m_rowStat[r].diagIntraSatd = 0; curEncData.m_rowStat[r].sumQpRc = 0; curEncData.m_rowStat[r].sumQpAq = 0; } m_bAllRowsStop = false; } } } /* SAO parameter estimation using non-deblocked pixels for CTU bottom and right boundary areas */ if (m_param->bEnableSAO && m_param->bSaoNonDeblocked) m_frameFilter.m_sao.calcSaoStatsCu_BeforeDblk(m_frame, col, row); if (m_param->bEnableWavefront && curRow.completed >= 2 && row < m_numRows - 1 && (!m_bAllRowsStop || intRow + 1 < m_vbvResetTriggerRow)) { /* activate next row */ ScopedLock below(m_rows[row + 1].lock); if (m_rows[row + 1].active == false && m_rows[row + 1].completed + 2 <= curRow.completed) { m_rows[row + 1].active = true; enqueueRowEncoder(row + 1); tryWakeOne(); /* wake up a sleeping thread or set the help wanted flag */ } } ScopedLock self(curRow.lock); if ((m_bAllRowsStop && intRow > m_vbvResetTriggerRow) || (row > 0 && curRow.completed < numCols - 1 && m_rows[row - 1].completed < m_rows[row].completed + 2)) { curRow.active = false; curRow.busy = false; ATOMIC_INC(&m_countRowBlocks); return; } } /** this row of CTUs has been compressed **/ /* If encoding with ABR, update update bits and complexity in rate control * after a number of rows so the next frame's rateControlStart has more * accurate data for estimation. At the start of the encode we update stats * after half the frame is encoded, but after this initial period we update * after refLagRows (the number of rows reference frames must have completed * before referencees may begin encoding) */ uint32_t rowCount = 0; if (m_param->rc.rateControlMode == X265_RC_ABR || bIsVbv) { if ((uint32_t)m_rce.encodeOrder <= 2 * (m_param->fpsNum / m_param->fpsDenom)) rowCount = X265_MIN((m_numRows + 1) / 2, m_numRows - 1); else rowCount = X265_MIN(m_refLagRows, m_numRows - 1); if (row == rowCount) { m_rce.rowTotalBits = 0; if (bIsVbv) for (uint32_t i = 0; i < rowCount; i++) m_rce.rowTotalBits += curEncData.m_rowStat[i].encodedBits; else for (uint32_t cuAddr = 0; cuAddr < rowCount * numCols; cuAddr++) m_rce.rowTotalBits += curEncData.m_cuStat[cuAddr].totalBits; m_top->m_rateControl->rateControlUpdateStats(&m_rce); } } /* flush row bitstream (if WPP and no SAO) or flush frame if no WPP and no SAO */ if (!m_param->bEnableSAO && (m_param->bEnableWavefront || row == m_numRows - 1)) rowCoder.finishSlice(); if (m_param->bEnableWavefront) { /* trigger row-wise loop filters */ if (row >= m_filterRowDelay) { enableRowFilter(row - m_filterRowDelay); /* NOTE: Activate filter if first row (row 0) */ if (row == m_filterRowDelay) enqueueRowFilter(0); tryWakeOne(); } if (row == m_numRows - 1) { for (uint32_t i = m_numRows - m_filterRowDelay; i < m_numRows; i++) enableRowFilter(i); tryWakeOne(); } } tld.analysis.m_param = NULL; curRow.busy = false; if (ATOMIC_INC(&m_completionCount) == 2 * (int)m_numRows) m_completionEvent.trigger();}/* collect statistics about CU coding decisions, return total QP */int FrameEncoder::collectCTUStatistics(const CUData& ctu, uint32_t* qtreeInterCnt, uint32_t* qtreeIntraCnt, uint32_t* qtreeSkipCnt){ StatisticLog* log = &m_sliceTypeLog[ctu.m_slice->m_sliceType]; int totQP = 0; if (ctu.m_slice->m_sliceType == I_SLICE) { uint32_t depth = 0; for (uint32_t absPartIdx = 0; absPartIdx < ctu.m_numPartitions; absPartIdx += ctu.m_numPartitions >> (depth * 2)) { depth = ctu.m_cuDepth[absPartIdx]; log->totalCu++; log->cntIntra[depth]++; qtreeIntraCnt[depth]++; totQP += ctu.m_qp[absPartIdx] * (ctu.m_numPartitions >> (depth * 2)); if (ctu.m_predMode[absPartIdx] == MODE_NONE) { log->totalCu--; log->cntIntra[depth]--; qtreeIntraCnt[depth]--; } else if (ctu.m_partSize[absPartIdx] != SIZE_2Nx2N) { /* TODO: log intra modes at absPartIdx +0 to +3 */ X265_CHECK(ctu.m_log2CUSize[absPartIdx] == 3 && ctu.m_slice->m_sps->quadtreeTULog2MinSize < 3, "Intra NxN found at improbable depth\n"); log->cntIntraNxN++; log->cntIntra[depth]--; } else if (ctu.m_lumaIntraDir[absPartIdx] > 1) log->cuIntraDistribution[depth][ANGULAR_MODE_ID]++; else log->cuIntraDistribution[depth][ctu.m_lumaIntraDir[absPartIdx]]++; } } else { uint32_t depth = 0; for (uint32_t absPartIdx = 0; absPartIdx < ctu.m_numPartitions; absPartIdx += ctu.m_numPartitions >> (depth * 2)) { depth = ctu.m_cuDepth[absPartIdx]; log->totalCu++; log->cntTotalCu[depth]++; totQP += ctu.m_qp[absPartIdx] * (ctu.m_numPartitions >> (depth * 2)); if (ctu.m_predMode[absPartIdx] == MODE_NONE) { log->totalCu--; log->cntTotalCu[depth]--; } else if (ctu.isSkipped(absPartIdx)) { log->totalCu--; log->cntSkipCu[depth]++; qtreeSkipCnt[depth]++; } else if (ctu.isInter(absPartIdx)) { log->cntInter[depth]++; qtreeInterCnt[depth]++; if (ctu.m_partSize[absPartIdx] < AMP_ID) log->cuInterDistribution[depth][ctu.m_partSize[absPartIdx]]++; else log->cuInterDistribution[depth][AMP_ID]++; } else if (ctu.isIntra(absPartIdx)) { log->cntIntra[depth]++; qtreeIntraCnt[depth]++; if (ctu.m_partSize[absPartIdx] != SIZE_2Nx2N) { X265_CHECK(ctu.m_log2CUSize[absPartIdx] == 3 && ctu.m_slice->m_sps->quadtreeTULog2MinSize < 3, "Intra NxN found at improbable depth\n"); log->cntIntraNxN++; log->cntIntra[depth]--; /* TODO: log intra modes at absPartIdx +0 to +3 */ } else if (ctu.m_lumaIntraDir[absPartIdx] > 1) log->cuIntraDistribution[depth][ANGULAR_MODE_ID]++; else log->cuIntraDistribution[depth][ctu.m_lumaIntraDir[absPartIdx]]++; } } } return totQP;}/* iterate over coded CUs and determine total QP */int FrameEncoder::calcCTUQP(const CUData& ctu){ int totQP = 0; uint32_t depth = 0, numParts = ctu.m_numPartitions; for (uint32_t absPartIdx = 0; absPartIdx < ctu.m_numPartitions; absPartIdx += numParts) { depth = ctu.m_cuDepth[absPartIdx]; numParts = ctu.m_numPartitions >> (depth * 2); totQP += ctu.m_qp[absPartIdx] * numParts; } return totQP;}/* DCT-domain noise reduction / adaptive deadzone from libavcodec */void FrameEncoder::noiseReductionUpdate(){ static const uint32_t maxBlocksPerTrSize[4] = {1 << 18, 1 << 16, 1 << 14, 1 << 12}; for (int cat = 0; cat < MAX_NUM_TR_CATEGORIES; cat++) { int trSize = cat & 3; int coefCount = 1 << ((trSize + 2) * 2); if (m_nr->count[cat] > maxBlocksPerTrSize[trSize]) { for (int i = 0; i < coefCount; i++) m_nr->residualSum[cat][i] >>= 1; m_nr->count[cat] >>= 1; } int nrStrength = cat < 8 ? m_param->noiseReductionIntra : m_param->noiseReductionInter; uint64_t scaledCount = (uint64_t)nrStrength * m_nr->count[cat]; for (int i = 0; i < coefCount; i++) { uint64_t value = scaledCount + m_nr->residualSum[cat][i] / 2; uint64_t denom = m_nr->residualSum[cat][i] + 1; m_nr->offsetDenoise[cat][i] = (uint16_t)(value / denom); } // Don't denoise DC coefficients m_nr->offsetDenoise[cat][0] = 0; }}Frame *FrameEncoder::getEncodedPicture(NALList& output){ if (m_frame)//m_done一开始为wait ,待初始化完毕触发一次 但是此时m_frame为null 因为m_frame 在m_enable.trigger()前为null { /* block here until worker thread completes */ m_done.wait();//执行wait操作,等待编码完毕 Frame *ret = m_frame; m_frame = NULL; output.takeContents(m_nalList); m_prevOutputTime = x265_mdate(); return ret; } return NULL;}}

 

转载地址:http://touub.baihongyu.com/

你可能感兴趣的文章
剑指offer 21.包含min函数的栈
查看>>
剑指offer 23.从上往下打印二叉树
查看>>
剑指offer 25.二叉树中和为某一值的路径
查看>>
剑指offer 60. 不用加减乘除做加法
查看>>
Leetcode C++《热题 Hot 100-14》283.移动零
查看>>
Leetcode C++《热题 Hot 100-15》437.路径总和III
查看>>
Leetcode C++《热题 Hot 100-17》461.汉明距离
查看>>
Leetcode C++《热题 Hot 100-18》538.把二叉搜索树转换为累加树
查看>>
Leetcode C++《热题 Hot 100-21》581.最短无序连续子数组
查看>>
Leetcode C++《热题 Hot 100-22》2.两数相加
查看>>
Leetcode C++《热题 Hot 100-23》3.无重复字符的最长子串
查看>>
Leetcode C++《热题 Hot 100-24》5.最长回文子串
查看>>
Leetcode C++《热题 Hot 100-28》19.删除链表的倒数第N个节点
查看>>
Leetcode C++《热题 Hot 100-29》22.括号生成
查看>>
Leetcode C++《热题 Hot 100-47》236.二叉树的最近公共祖先
查看>>
Leetcode C++《热题 Hot 100-48》406.根据身高重建队列
查看>>
《kubernetes权威指南·第四版》第二章:kubernetes安装配置指南
查看>>
Leetcode C++《热题 Hot 100-49》399.除法求值
查看>>
Leetcode C++《热题 Hot 100-51》152. 乘积最大子序列
查看>>
Leetcode C++ 《第181场周赛-1》 5364. 按既定顺序创建目标数组
查看>>