@@ -17,16 +17,6 @@ cmpc::CMpegDecoder::CMpegDecoder(void)
17
17
videoPath.clear ();
18
18
_str_codec.clear ();
19
19
20
- for (auto i = std::begin (video_dst_data); i < std::end (video_dst_data); i++) {
21
- *i = nullptr ;
22
- }
23
-
24
- for (auto i = std::begin (video_dst_linesize); i < std::end (video_dst_linesize); i++) {
25
- *i = 0 ;
26
- }
27
-
28
- video_dst_bufsize = 0 ;
29
-
30
20
/* Enable or disable frame reference counting. You are not supposed to support
31
21
* both paths in your application but pick the one most appropriate to your
32
22
* needs. Look for the use of refcount in this example to see what are the
@@ -59,34 +49,18 @@ void cmpc::CMpegDecoder::clear(void) {
59
49
sws_freeContext (PswsCtx);
60
50
PswsCtx = nullptr ;
61
51
}
62
- // PswsCtx = nullptr;
63
- // cout << "SWS Freed!" << endl;
64
52
if (RGBbuffer) {
65
53
av_free (RGBbuffer);
66
54
RGBbuffer = nullptr ;
67
55
}
68
- // cout << "Buffer Freed!" << endl;
69
56
if (PCodecCtx) {
70
57
avcodec_free_context (&PCodecCtx);
71
58
PCodecCtx = nullptr ;
72
59
}
73
- // cout << "Condec Contex Freed!" << endl;
74
60
if (PFormatCtx) {
75
61
avformat_close_input (&PFormatCtx);
76
62
PFormatCtx = nullptr ;
77
63
}
78
- // cout << "Format Contex Freed!" << endl;
79
- if (video_dst_data[0 ])
80
- av_freep (&video_dst_data[0 ]);
81
- // cout << "DST Buffer Freed!" << endl;
82
- for (auto i = std::begin (video_dst_data); i < std::end (video_dst_data); i++) {
83
- *i = nullptr ;
84
- }
85
- for (auto i = std::begin (video_dst_linesize); i < std::end (video_dst_linesize); i++) {
86
- *i = 0 ;
87
- }
88
-
89
- video_dst_bufsize = 0 ;
90
64
refcount = 1 ;
91
65
}
92
66
@@ -120,14 +94,6 @@ cmpc::CMpegDecoder::CMpegDecoder(CMpegDecoder &&ref) noexcept
120
94
ref.PCodecCtx = nullptr ;
121
95
ref.PVideoStream = nullptr ;
122
96
ref.PswsCtx = nullptr ;
123
- for (auto i = std::begin (video_dst_data), j = std::begin (ref.video_dst_data ); \
124
- i < std::end (video_dst_data), j = std::end (ref.video_dst_data ); i++, j++) {
125
- *i = *j;
126
- }
127
- for (auto i = std::begin (video_dst_linesize), j = std::begin (ref.video_dst_linesize ); \
128
- i < std::end (video_dst_linesize), j = std::end (ref.video_dst_linesize ); i++, j++) {
129
- *i = *j;
130
- }
131
97
}
132
98
133
99
cmpc::CMpegDecoder& cmpc::CMpegDecoder::operator =(CMpegDecoder &&ref) noexcept {
@@ -155,14 +121,6 @@ cmpc::CMpegDecoder& cmpc::CMpegDecoder::operator=(CMpegDecoder &&ref) noexcept {
155
121
ref.PVideoStream = nullptr ;
156
122
ref.RGBbuffer = nullptr ;
157
123
ref.PswsCtx = nullptr ;
158
- for (auto i = std::begin (video_dst_data), j = std::begin (ref.video_dst_data ); \
159
- i < std::end (video_dst_data), j = std::end (ref.video_dst_data ); i++, j++) {
160
- *i = *j;
161
- }
162
- for (auto i = std::begin (video_dst_linesize), j = std::begin (ref.video_dst_linesize ); \
163
- i < std::end (video_dst_linesize), j = std::end (ref.video_dst_linesize ); i++, j++) {
164
- *i = *j;
165
- }
166
124
refcount = ref.refcount ;
167
125
}
168
126
return *this ;
@@ -268,14 +226,6 @@ bool cmpc::CMpegDecoder::FFmpegSetup() { //打开指定路径的视频文件,
268
226
PPixelFormat = PCodecCtx->pix_fmt ;
269
227
_duration = static_cast <double >(PVideoStream->duration ) / static_cast <double >(time_base.den ) * static_cast <double >(time_base.num );
270
228
_predictFrameNum = av_rescale (static_cast <int64_t >(_duration*0xFFFF ), frame_base.num , frame_base.den )/0xFFFF ;
271
- ret = av_image_alloc (video_dst_data, video_dst_linesize,
272
- width, height, PPixelFormat, 1 ); // 使用源格式的,一律保持原状
273
- if (ret < 0 ) {
274
- cerr << " Could not allocate raw video buffer" << endl;
275
- clear ();
276
- return false ;
277
- }
278
- video_dst_bufsize = ret;
279
229
}
280
230
281
231
/* dump input information to stderr */
@@ -292,13 +242,11 @@ bool cmpc::CMpegDecoder::FFmpegSetup() { //打开指定路径的视频文件,
292
242
if (widthDst > 0 && heightDst > 0 ) {
293
243
PswsCtx = sws_getContext (width, height, PPixelFormat, widthDst, heightDst, AV_PIX_FMT_RGB24, SCALE_FLAGS, nullptr , nullptr , nullptr );
294
244
auto numBytes = av_image_get_buffer_size (AV_PIX_FMT_RGB24, widthDst, heightDst, 1 );
295
- video_dst_bufsize = numBytes;
296
245
RGBbuffer = (uint8_t *)av_malloc (numBytes * sizeof (uint8_t ));
297
246
}
298
247
else {
299
248
PswsCtx = sws_getContext (width, height, PPixelFormat, width, height, AV_PIX_FMT_RGB24, SCALE_FLAGS, nullptr , nullptr , nullptr );
300
249
auto numBytes = av_image_get_buffer_size (AV_PIX_FMT_RGB24, width, height, 1 );
301
- video_dst_bufsize = numBytes;
302
250
RGBbuffer = (uint8_t *)av_malloc (numBytes * sizeof (uint8_t ));
303
251
}
304
252
return true ;
@@ -427,22 +375,15 @@ int cmpc::CMpegDecoder::_SaveFrame(PyObject *PyFrameList, AVFrame *&frame, AVFra
427
375
428
376
/* copy decoded frame to destination buffer:
429
377
* this is required since rawvideo expects non aligned data */
430
- /* av_image_copy(video_dst_data, video_dst_linesize,
431
- (const uint8_t **)frame->data, frame->linesize,
432
- PPixelFormat, width, height);*/
433
378
434
379
sws_scale (PswsCtx, frame->data , frame->linesize , 0 , height, frameRGB->data , frameRGB->linesize );
435
-
436
- // cout << "Complete Conv ";
437
380
438
381
/* write to rawvideo file */
439
382
if (widthDst>0 && heightDst>0 )
440
383
OneFrame = _SaveFrame_castToPyFrameArray (frameRGB->data , widthDst, heightDst);
441
384
else
442
385
OneFrame = _SaveFrame_castToPyFrameArray (frameRGB->data , width, height);
443
386
PyList_Append (PyFrameList, OneFrame);
444
- // cout << "[" << width << "-" << height << ", " << width*height << ", " << video_dst_bufsize << "]" << endl;
445
- // cout << "PTS = " << frameRGB->pts << ", coded Fnum = " << frameRGB->coded_picture_number << endl;
446
387
processed = true ;
447
388
}
448
389
}
@@ -523,8 +464,6 @@ int cmpc::CMpegDecoder::_SaveFrameForGOP(PyObject *PyFrameList, AVFrame *&frame,
523
464
524
465
sws_scale (PswsCtx, frame->data , frame->linesize , 0 , height, frameRGB->data , frameRGB->linesize );
525
466
526
- // cout << "Complete Conv ";
527
-
528
467
/* write to rawvideo file */
529
468
if (widthDst>0 && heightDst>0 )
530
469
OneFrame = _SaveFrame_castToPyFrameArray (frameRGB->data , widthDst, heightDst);
@@ -768,7 +707,6 @@ bool cmpc::CMpegDecoder::ExtractFrame(PyObject* PyFrameList, int64_t framePos, i
768
707
else {
769
708
framePos_TimeBase = __FrameToPts (framePos);
770
709
}
771
- // cout << framePos_TimeBase << endl;
772
710
if (av_seek_frame (PFormatCtx, PVideoStreamIDX, framePos_TimeBase, AVSEEK_FLAG_BACKWARD) < 0 ) {
773
711
cerr << " AV seek frame fail!" << endl;
774
712
av_seek_frame (PFormatCtx, -1 , 0 , AVSEEK_FLAG_BACKWARD);
@@ -783,7 +721,6 @@ bool cmpc::CMpegDecoder::ExtractFrame(PyObject* PyFrameList, int64_t framePos, i
783
721
}
784
722
785
723
while (av_read_frame (PFormatCtx, &pkt) >= 0 ) {
786
- // cout << "[Test - " << pkt.size << " ]" << endl;
787
724
AVPacket orig_pkt = pkt;
788
725
frameProcessed = false ;
789
726
do {
@@ -807,23 +744,13 @@ bool cmpc::CMpegDecoder::ExtractFrame(PyObject* PyFrameList, int64_t framePos, i
807
744
_SaveFrame (PyFrameList, frame, frameRGB, pkt, got_frame, framePos_TimeBase, frameProcessed, 1 );
808
745
} while (got_frame);
809
746
810
- // cout << "Demuxing succeeded." << endl;
811
-
812
747
if (PVideoStream && count>0 && (__dumpControl > 0 )) {
813
748
cout << " Succeed in convert frames into Python_List" << endl;
814
749
}
815
750
816
- // av_free(RGBbuffer);
817
- // RGBbuffer = nullptr;
818
- // cout << "Free Buffer" << endl;
819
- // sws_freeContext(PswsCtx);
820
- // cout << "Free ctx" << endl;
821
- // PswsCtx = nullptr;
822
751
av_frame_free (&frameRGB);
823
752
av_frame_free (&frame);
824
753
825
- // cout << "End Process" << endl;
826
-
827
754
return true ;
828
755
}
829
756
@@ -925,7 +852,7 @@ cmpc::CMpegEncoder::CMpegEncoder(CMpegEncoder &&ref) noexcept:
925
852
codecName.assign (ref.codecName );
926
853
}
927
854
928
- cmpc::CMpegEncoder& cmpc::CMpegEncoder::operator =(CMpegEncoder &&ref) noexcept {
855
+ cmpc::CMpegEncoder& cmpc::CMpegEncoder::operator =(CMpegEncoder &&ref) noexcept {
929
856
videoPath.assign (ref.videoPath );
930
857
codecName.assign (ref.codecName );
931
858
bitRate = ref.bitRate ;
@@ -981,8 +908,6 @@ void cmpc::CMpegEncoder::__log_packet(){
981
908
AVRational *time_base = &PFormatCtx->streams [Ppacket->stream_index ]->time_base ;
982
909
cout << " pts:" << av_ts2str (Ppacket->pts ) << " pts_time:" << av_ts2timestr (Ppacket->pts , time_base)
983
910
<< " dts:" << av_ts2str (Ppacket->dts ) << " dts_time:" << av_ts2timestr (Ppacket->dts , time_base) << endl;
984
- // << " duration:" << av_ts2str(Ppacket->duration) << " duration_time:"
985
- // << av_ts2timestr(Ppacket->duration, time_base) << " stream_index:"<< Ppacket->stream_index << endl;
986
911
}
987
912
988
913
int cmpc::CMpegEncoder::__write_frame (){
@@ -1012,8 +937,6 @@ bool cmpc::CMpegEncoder::__add_stream(AVCodec **codec){
1012
937
codec_id = (*codec)->id ;
1013
938
PFormatCtx->oformat ->video_codec = codec_id;
1014
939
}
1015
- // auto codec_id = PFormatCtx->oformat->video_codec;
1016
- // *codec = avcodec_find_encoder(codec_id);
1017
940
if (!(*codec)) {
1018
941
cerr << " Could not find encoder for '" << avcodec_get_name (codec_id) << " '" << endl;
1019
942
return false ;
@@ -1044,11 +967,13 @@ bool cmpc::CMpegEncoder::__add_stream(AVCodec **codec){
1044
967
* of which frame timestamps are represented. For fixed-fps content,
1045
968
* timebase should be 1/framerate and timestamp increments should be
1046
969
* identical to 1. */
1047
- PStreamContex.st ->time_base = timeBase;
970
+ PStreamContex.st ->time_base .den = 0 ;
971
+ PStreamContex.st ->time_base .num = 0 ;
1048
972
// av_stream_set_r_frame_rate(PStreamContex.st, frameRate);
1049
973
// cout << "(" << frameRate.num << ", " << frameRate.den << ")" << endl;
1050
974
// PStreamContex.st->r_frame_rate
1051
- c->time_base = PStreamContex.st ->time_base ;
975
+ c->time_base = timeBase;
976
+
1052
977
// PStreamContex.st->frame
1053
978
c->framerate = frameRate;
1054
979
@@ -1204,11 +1129,8 @@ AVFrame *cmpc::CMpegEncoder::__get_video_frame(PyArrayObject* PyFrame) {
1204
1129
}
1205
1130
}
1206
1131
1207
- // PStreamContex.next_frame++;
1208
- // PStreamContex.frame->pts = __FrameToPts(PStreamContex.next_frame);
1209
1132
PStreamContex.frame ->pts = PStreamContex.next_frame ;
1210
1133
PStreamContex.next_frame ++;
1211
- // PStreamContex.frame->pts = PStreamContex.next_pts++;
1212
1134
return PStreamContex.frame ;
1213
1135
}
1214
1136
@@ -1439,14 +1361,6 @@ void cmpc::CMpegEncoder::setParameter(string keyword, void *ptr) {
1439
1361
int *ref = reinterpret_cast <int *>(ptr);
1440
1362
MaxBFrame = *ref;
1441
1363
}
1442
- /* else if (keyword.compare("timeBase") == 0) {
1443
- PyObject *ref = reinterpret_cast<PyObject *>(ptr);
1444
- auto refObj = PyTuple_GetItem(ref, 0);
1445
- int num = static_cast<int>(PyLong_AsLong(refObj));
1446
- refObj = PyTuple_GetItem(ref, 1);
1447
- int den = static_cast<int>(PyLong_AsLong(refObj));
1448
- timeBase = _setAVRational(num, den);
1449
- }*/
1450
1364
else if (keyword.compare (" frameRate" ) == 0 ) {
1451
1365
PyObject *ref = reinterpret_cast <PyObject *>(ptr);
1452
1366
auto refObj = PyTuple_GetItem (ref, 0 );
@@ -1476,7 +1390,7 @@ bool cmpc::CMpegEncoder::FFmpegSetup() {
1476
1390
/* allocate the output media context */
1477
1391
// auto getFormat = av_guess_format(codecName.c_str(), nullptr, nullptr);
1478
1392
avformat_alloc_output_context2 (&PFormatCtx, nullptr , nullptr , videoPath.c_str ());
1479
- PFormatCtx->avoid_negative_ts = AVFMT_AVOID_NEG_TS_MAKE_ZERO ;
1393
+ PFormatCtx->avoid_negative_ts = AVFMT_AVOID_NEG_TS_AUTO ;
1480
1394
if (!PFormatCtx) {
1481
1395
cout << " Could not select the encoder automatically: using MPEG." << endl;
1482
1396
// cout << "Could not deduce output format from file extension: using MPEG." << endl;
0 commit comments