avconv: use avcodec_encode_audio2()
[libav.git] / avconv.c
1 /*
2 * avconv main
3 * Copyright (c) 2000-2011 The libav developers.
4 *
5 * This file is part of Libav.
6 *
7 * Libav is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
11 *
12 * Libav is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with Libav; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20 */
21
22 #include "config.h"
23 #include <ctype.h>
24 #include <string.h>
25 #include <math.h>
26 #include <stdlib.h>
27 #include <errno.h>
28 #include <signal.h>
29 #include <limits.h>
30 #include <unistd.h>
31 #include "libavformat/avformat.h"
32 #include "libavdevice/avdevice.h"
33 #include "libswscale/swscale.h"
34 #include "libavutil/opt.h"
35 #include "libavcodec/audioconvert.h"
36 #include "libavutil/audioconvert.h"
37 #include "libavutil/parseutils.h"
38 #include "libavutil/samplefmt.h"
39 #include "libavutil/colorspace.h"
40 #include "libavutil/fifo.h"
41 #include "libavutil/intreadwrite.h"
42 #include "libavutil/dict.h"
43 #include "libavutil/mathematics.h"
44 #include "libavutil/pixdesc.h"
45 #include "libavutil/avstring.h"
46 #include "libavutil/libm.h"
47 #include "libavutil/imgutils.h"
48 #include "libavformat/os_support.h"
49
50 #if CONFIG_AVFILTER
51 # include "libavfilter/avfilter.h"
52 # include "libavfilter/avfiltergraph.h"
53 # include "libavfilter/buffersrc.h"
54 # include "libavfilter/vsrc_buffer.h"
55 #endif
56
57 #if HAVE_SYS_RESOURCE_H
58 #include <sys/types.h>
59 #include <sys/time.h>
60 #include <sys/resource.h>
61 #elif HAVE_GETPROCESSTIMES
62 #include <windows.h>
63 #endif
64 #if HAVE_GETPROCESSMEMORYINFO
65 #include <windows.h>
66 #include <psapi.h>
67 #endif
68
69 #if HAVE_SYS_SELECT_H
70 #include <sys/select.h>
71 #endif
72
73 #include <time.h>
74
75 #include "cmdutils.h"
76
77 #include "libavutil/avassert.h"
78
79 #define VSYNC_AUTO -1
80 #define VSYNC_PASSTHROUGH 0
81 #define VSYNC_CFR 1
82 #define VSYNC_VFR 2
83
84 const char program_name[] = "avconv";
85 const int program_birth_year = 2000;
86
87 /* select an input stream for an output stream */
88 typedef struct StreamMap {
89 int disabled; /** 1 is this mapping is disabled by a negative map */
90 int file_index;
91 int stream_index;
92 int sync_file_index;
93 int sync_stream_index;
94 } StreamMap;
95
96 /**
97 * select an input file for an output file
98 */
99 typedef struct MetadataMap {
100 int file; ///< file index
101 char type; ///< type of metadata to copy -- (g)lobal, (s)tream, (c)hapter or (p)rogram
102 int index; ///< stream/chapter/program number
103 } MetadataMap;
104
105 static const OptionDef options[];
106
107 static int video_discard = 0;
108 static int same_quant = 0;
109 static int do_deinterlace = 0;
110 static int intra_dc_precision = 8;
111 static int qp_hist = 0;
112
113 static int file_overwrite = 0;
114 static int do_benchmark = 0;
115 static int do_hex_dump = 0;
116 static int do_pkt_dump = 0;
117 static int do_pass = 0;
118 static char *pass_logfilename_prefix = NULL;
119 static int video_sync_method = VSYNC_AUTO;
120 static int audio_sync_method = 0;
121 static float audio_drift_threshold = 0.1;
122 static int copy_ts = 0;
123 static int copy_tb = 1;
124 static int opt_shortest = 0;
125 static char *vstats_filename;
126 static FILE *vstats_file;
127
128 static int audio_volume = 256;
129
130 static int exit_on_error = 0;
131 static int using_stdin = 0;
132 static int64_t video_size = 0;
133 static int64_t audio_size = 0;
134 static int64_t extra_size = 0;
135 static int nb_frames_dup = 0;
136 static int nb_frames_drop = 0;
137 static int input_sync;
138
139 static float dts_delta_threshold = 10;
140
141 static int print_stats = 1;
142
143 static uint8_t *audio_buf;
144 static unsigned int allocated_audio_buf_size;
145
146 #define DEFAULT_PASS_LOGFILENAME_PREFIX "av2pass"
147
148 typedef struct FrameBuffer {
149 uint8_t *base[4];
150 uint8_t *data[4];
151 int linesize[4];
152
153 int h, w;
154 enum PixelFormat pix_fmt;
155
156 int refcount;
157 struct InputStream *ist;
158 struct FrameBuffer *next;
159 } FrameBuffer;
160
161 typedef struct InputStream {
162 int file_index;
163 AVStream *st;
164 int discard; /* true if stream data should be discarded */
165 int decoding_needed; /* true if the packets must be decoded in 'raw_fifo' */
166 AVCodec *dec;
167 AVFrame *decoded_frame;
168 AVFrame *filtered_frame;
169
170 int64_t start; /* time when read started */
171 int64_t next_pts; /* synthetic pts for cases where pkt.pts
172 is not defined */
173 int64_t pts; /* current pts */
174 PtsCorrectionContext pts_ctx;
175 double ts_scale;
176 int is_start; /* is 1 at the start and after a discontinuity */
177 int showed_multi_packet_warning;
178 AVDictionary *opts;
179
180 /* a pool of free buffers for decoded data */
181 FrameBuffer *buffer_pool;
182 } InputStream;
183
184 typedef struct InputFile {
185 AVFormatContext *ctx;
186 int eof_reached; /* true if eof reached */
187 int ist_index; /* index of first stream in ist_table */
188 int buffer_size; /* current total buffer size */
189 int64_t ts_offset;
190 int nb_streams; /* number of stream that avconv is aware of; may be different
191 from ctx.nb_streams if new streams appear during av_read_frame() */
192 int rate_emu;
193 } InputFile;
194
195 typedef struct OutputStream {
196 int file_index; /* file index */
197 int index; /* stream index in the output file */
198 int source_index; /* InputStream index */
199 AVStream *st; /* stream in the output file */
200 int encoding_needed; /* true if encoding needed for this stream */
201 int frame_number;
202 /* input pts and corresponding output pts
203 for A/V sync */
204 // double sync_ipts; /* dts from the AVPacket of the demuxer in second units */
205 struct InputStream *sync_ist; /* input stream to sync against */
206 int64_t sync_opts; /* output frame counter, could be changed to some true timestamp */ // FIXME look at frame_number
207 AVBitStreamFilterContext *bitstream_filters;
208 AVCodec *enc;
209 int64_t max_frames;
210 AVFrame *output_frame;
211
212 /* video only */
213 int video_resample;
214 AVFrame pict_tmp; /* temporary image for resampling */
215 struct SwsContext *img_resample_ctx; /* for image resampling */
216 int resample_height;
217 int resample_width;
218 int resample_pix_fmt;
219 AVRational frame_rate;
220 int force_fps;
221 int top_field_first;
222
223 float frame_aspect_ratio;
224
225 /* forced key frames */
226 int64_t *forced_kf_pts;
227 int forced_kf_count;
228 int forced_kf_index;
229
230 /* audio only */
231 int audio_resample;
232 ReSampleContext *resample; /* for audio resampling */
233 int resample_sample_fmt;
234 int resample_channels;
235 int resample_sample_rate;
236 int reformat_pair;
237 AVAudioConvert *reformat_ctx;
238 AVFifoBuffer *fifo; /* for compression: one audio fifo per codec */
239 FILE *logfile;
240
241 #if CONFIG_AVFILTER
242 AVFilterContext *output_video_filter;
243 AVFilterContext *input_video_filter;
244 AVFilterBufferRef *picref;
245 char *avfilter;
246 AVFilterGraph *graph;
247 #endif
248
249 int64_t sws_flags;
250 AVDictionary *opts;
251 int is_past_recording_time;
252 int stream_copy;
253 const char *attachment_filename;
254 int copy_initial_nonkeyframes;
255 } OutputStream;
256
257
258 typedef struct OutputFile {
259 AVFormatContext *ctx;
260 AVDictionary *opts;
261 int ost_index; /* index of the first stream in output_streams */
262 int64_t recording_time; /* desired length of the resulting file in microseconds */
263 int64_t start_time; /* start time in microseconds */
264 uint64_t limit_filesize;
265 } OutputFile;
266
267 static InputStream *input_streams = NULL;
268 static int nb_input_streams = 0;
269 static InputFile *input_files = NULL;
270 static int nb_input_files = 0;
271
272 static OutputStream *output_streams = NULL;
273 static int nb_output_streams = 0;
274 static OutputFile *output_files = NULL;
275 static int nb_output_files = 0;
276
277 typedef struct OptionsContext {
278 /* input/output options */
279 int64_t start_time;
280 const char *format;
281
282 SpecifierOpt *codec_names;
283 int nb_codec_names;
284 SpecifierOpt *audio_channels;
285 int nb_audio_channels;
286 SpecifierOpt *audio_sample_rate;
287 int nb_audio_sample_rate;
288 SpecifierOpt *frame_rates;
289 int nb_frame_rates;
290 SpecifierOpt *frame_sizes;
291 int nb_frame_sizes;
292 SpecifierOpt *frame_pix_fmts;
293 int nb_frame_pix_fmts;
294
295 /* input options */
296 int64_t input_ts_offset;
297 int rate_emu;
298
299 SpecifierOpt *ts_scale;
300 int nb_ts_scale;
301 SpecifierOpt *dump_attachment;
302 int nb_dump_attachment;
303
304 /* output options */
305 StreamMap *stream_maps;
306 int nb_stream_maps;
307 /* first item specifies output metadata, second is input */
308 MetadataMap (*meta_data_maps)[2];
309 int nb_meta_data_maps;
310 int metadata_global_manual;
311 int metadata_streams_manual;
312 int metadata_chapters_manual;
313 const char **attachments;
314 int nb_attachments;
315
316 int chapters_input_file;
317
318 int64_t recording_time;
319 uint64_t limit_filesize;
320 float mux_preload;
321 float mux_max_delay;
322
323 int video_disable;
324 int audio_disable;
325 int subtitle_disable;
326 int data_disable;
327
328 /* indexed by output file stream index */
329 int *streamid_map;
330 int nb_streamid_map;
331
332 SpecifierOpt *metadata;
333 int nb_metadata;
334 SpecifierOpt *max_frames;
335 int nb_max_frames;
336 SpecifierOpt *bitstream_filters;
337 int nb_bitstream_filters;
338 SpecifierOpt *codec_tags;
339 int nb_codec_tags;
340 SpecifierOpt *sample_fmts;
341 int nb_sample_fmts;
342 SpecifierOpt *qscale;
343 int nb_qscale;
344 SpecifierOpt *forced_key_frames;
345 int nb_forced_key_frames;
346 SpecifierOpt *force_fps;
347 int nb_force_fps;
348 SpecifierOpt *frame_aspect_ratios;
349 int nb_frame_aspect_ratios;
350 SpecifierOpt *rc_overrides;
351 int nb_rc_overrides;
352 SpecifierOpt *intra_matrices;
353 int nb_intra_matrices;
354 SpecifierOpt *inter_matrices;
355 int nb_inter_matrices;
356 SpecifierOpt *top_field_first;
357 int nb_top_field_first;
358 SpecifierOpt *metadata_map;
359 int nb_metadata_map;
360 SpecifierOpt *presets;
361 int nb_presets;
362 SpecifierOpt *copy_initial_nonkeyframes;
363 int nb_copy_initial_nonkeyframes;
364 #if CONFIG_AVFILTER
365 SpecifierOpt *filters;
366 int nb_filters;
367 #endif
368 } OptionsContext;
369
370 #define MATCH_PER_STREAM_OPT(name, type, outvar, fmtctx, st)\
371 {\
372 int i, ret;\
373 for (i = 0; i < o->nb_ ## name; i++) {\
374 char *spec = o->name[i].specifier;\
375 if ((ret = check_stream_specifier(fmtctx, st, spec)) > 0)\
376 outvar = o->name[i].u.type;\
377 else if (ret < 0)\
378 exit_program(1);\
379 }\
380 }
381
382 static void reset_options(OptionsContext *o)
383 {
384 const OptionDef *po = options;
385
386 /* all OPT_SPEC and OPT_STRING can be freed in generic way */
387 while (po->name) {
388 void *dst = (uint8_t*)o + po->u.off;
389
390 if (po->flags & OPT_SPEC) {
391 SpecifierOpt **so = dst;
392 int i, *count = (int*)(so + 1);
393 for (i = 0; i < *count; i++) {
394 av_freep(&(*so)[i].specifier);
395 if (po->flags & OPT_STRING)
396 av_freep(&(*so)[i].u.str);
397 }
398 av_freep(so);
399 *count = 0;
400 } else if (po->flags & OPT_OFFSET && po->flags & OPT_STRING)
401 av_freep(dst);
402 po++;
403 }
404
405 av_freep(&o->stream_maps);
406 av_freep(&o->meta_data_maps);
407 av_freep(&o->streamid_map);
408
409 memset(o, 0, sizeof(*o));
410
411 o->mux_max_delay = 0.7;
412 o->recording_time = INT64_MAX;
413 o->limit_filesize = UINT64_MAX;
414 o->chapters_input_file = INT_MAX;
415
416 uninit_opts();
417 init_opts();
418 }
419
420 static int alloc_buffer(InputStream *ist, FrameBuffer **pbuf)
421 {
422 AVCodecContext *s = ist->st->codec;
423 FrameBuffer *buf = av_mallocz(sizeof(*buf));
424 int ret;
425 const int pixel_size = av_pix_fmt_descriptors[s->pix_fmt].comp[0].step_minus1+1;
426 int h_chroma_shift, v_chroma_shift;
427 int edge = 32; // XXX should be avcodec_get_edge_width(), but that fails on svq1
428 int w = s->width, h = s->height;
429
430 if (!buf)
431 return AVERROR(ENOMEM);
432
433 if (!(s->flags & CODEC_FLAG_EMU_EDGE)) {
434 w += 2*edge;
435 h += 2*edge;
436 }
437
438 avcodec_align_dimensions(s, &w, &h);
439 if ((ret = av_image_alloc(buf->base, buf->linesize, w, h,
440 s->pix_fmt, 32)) < 0) {
441 av_freep(&buf);
442 return ret;
443 }
444 /* XXX this shouldn't be needed, but some tests break without this line
445 * those decoders are buggy and need to be fixed.
446 * the following tests fail:
447 * bethsoft-vid, cdgraphics, ansi, aasc, fraps-v1, qtrle-1bit
448 */
449 memset(buf->base[0], 128, ret);
450
451 avcodec_get_chroma_sub_sample(s->pix_fmt, &h_chroma_shift, &v_chroma_shift);
452 for (int i = 0; i < FF_ARRAY_ELEMS(buf->data); i++) {
453 const int h_shift = i==0 ? 0 : h_chroma_shift;
454 const int v_shift = i==0 ? 0 : v_chroma_shift;
455 if (s->flags & CODEC_FLAG_EMU_EDGE)
456 buf->data[i] = buf->base[i];
457 else
458 buf->data[i] = buf->base[i] +
459 FFALIGN((buf->linesize[i]*edge >> v_shift) +
460 (pixel_size*edge >> h_shift), 32);
461 }
462 buf->w = s->width;
463 buf->h = s->height;
464 buf->pix_fmt = s->pix_fmt;
465 buf->ist = ist;
466
467 *pbuf = buf;
468 return 0;
469 }
470
471 static void free_buffer_pool(InputStream *ist)
472 {
473 FrameBuffer *buf = ist->buffer_pool;
474 while (buf) {
475 ist->buffer_pool = buf->next;
476 av_freep(&buf->base[0]);
477 av_free(buf);
478 buf = ist->buffer_pool;
479 }
480 }
481
482 static void unref_buffer(InputStream *ist, FrameBuffer *buf)
483 {
484 av_assert0(buf->refcount);
485 buf->refcount--;
486 if (!buf->refcount) {
487 buf->next = ist->buffer_pool;
488 ist->buffer_pool = buf;
489 }
490 }
491
492 static int codec_get_buffer(AVCodecContext *s, AVFrame *frame)
493 {
494 InputStream *ist = s->opaque;
495 FrameBuffer *buf;
496 int ret, i;
497
498 if (!ist->buffer_pool && (ret = alloc_buffer(ist, &ist->buffer_pool)) < 0)
499 return ret;
500
501 buf = ist->buffer_pool;
502 ist->buffer_pool = buf->next;
503 buf->next = NULL;
504 if (buf->w != s->width || buf->h != s->height || buf->pix_fmt != s->pix_fmt) {
505 av_freep(&buf->base[0]);
506 av_free(buf);
507 if ((ret = alloc_buffer(ist, &buf)) < 0)
508 return ret;
509 }
510 buf->refcount++;
511
512 frame->opaque = buf;
513 frame->type = FF_BUFFER_TYPE_USER;
514 frame->extended_data = frame->data;
515 frame->pkt_pts = s->pkt ? s->pkt->pts : AV_NOPTS_VALUE;
516
517 for (i = 0; i < FF_ARRAY_ELEMS(buf->data); i++) {
518 frame->base[i] = buf->base[i]; // XXX h264.c uses base though it shouldn't
519 frame->data[i] = buf->data[i];
520 frame->linesize[i] = buf->linesize[i];
521 }
522
523 return 0;
524 }
525
526 static void codec_release_buffer(AVCodecContext *s, AVFrame *frame)
527 {
528 InputStream *ist = s->opaque;
529 FrameBuffer *buf = frame->opaque;
530 int i;
531
532 for (i = 0; i < FF_ARRAY_ELEMS(frame->data); i++)
533 frame->data[i] = NULL;
534
535 unref_buffer(ist, buf);
536 }
537
538 static void filter_release_buffer(AVFilterBuffer *fb)
539 {
540 FrameBuffer *buf = fb->priv;
541 av_free(fb);
542 unref_buffer(buf->ist, buf);
543 }
544
545 #if CONFIG_AVFILTER
546
547 static int configure_video_filters(InputStream *ist, OutputStream *ost)
548 {
549 AVFilterContext *last_filter, *filter;
550 /** filter graph containing all filters including input & output */
551 AVCodecContext *codec = ost->st->codec;
552 AVCodecContext *icodec = ist->st->codec;
553 FFSinkContext ffsink_ctx = { .pix_fmt = codec->pix_fmt };
554 AVRational sample_aspect_ratio;
555 char args[255];
556 int ret;
557
558 ost->graph = avfilter_graph_alloc();
559
560 if (ist->st->sample_aspect_ratio.num) {
561 sample_aspect_ratio = ist->st->sample_aspect_ratio;
562 } else
563 sample_aspect_ratio = ist->st->codec->sample_aspect_ratio;
564
565 snprintf(args, 255, "%d:%d:%d:%d:%d:%d:%d", ist->st->codec->width,
566 ist->st->codec->height, ist->st->codec->pix_fmt, 1, AV_TIME_BASE,
567 sample_aspect_ratio.num, sample_aspect_ratio.den);
568
569 ret = avfilter_graph_create_filter(&ost->input_video_filter, avfilter_get_by_name("buffer"),
570 "src", args, NULL, ost->graph);
571 if (ret < 0)
572 return ret;
573 ret = avfilter_graph_create_filter(&ost->output_video_filter, &ffsink,
574 "out", NULL, &ffsink_ctx, ost->graph);
575 if (ret < 0)
576 return ret;
577 last_filter = ost->input_video_filter;
578
579 if (codec->width != icodec->width || codec->height != icodec->height) {
580 snprintf(args, 255, "%d:%d:flags=0x%X",
581 codec->width,
582 codec->height,
583 (unsigned)ost->sws_flags);
584 if ((ret = avfilter_graph_create_filter(&filter, avfilter_get_by_name("scale"),
585 NULL, args, NULL, ost->graph)) < 0)
586 return ret;
587 if ((ret = avfilter_link(last_filter, 0, filter, 0)) < 0)
588 return ret;
589 last_filter = filter;
590 }
591
592 snprintf(args, sizeof(args), "flags=0x%X", (unsigned)ost->sws_flags);
593 ost->graph->scale_sws_opts = av_strdup(args);
594
595 if (ost->avfilter) {
596 AVFilterInOut *outputs = av_malloc(sizeof(AVFilterInOut));
597 AVFilterInOut *inputs = av_malloc(sizeof(AVFilterInOut));
598
599 outputs->name = av_strdup("in");
600 outputs->filter_ctx = last_filter;
601 outputs->pad_idx = 0;
602 outputs->next = NULL;
603
604 inputs->name = av_strdup("out");
605 inputs->filter_ctx = ost->output_video_filter;
606 inputs->pad_idx = 0;
607 inputs->next = NULL;
608
609 if ((ret = avfilter_graph_parse(ost->graph, ost->avfilter, inputs, outputs, NULL)) < 0)
610 return ret;
611 } else {
612 if ((ret = avfilter_link(last_filter, 0, ost->output_video_filter, 0)) < 0)
613 return ret;
614 }
615
616 if ((ret = avfilter_graph_config(ost->graph, NULL)) < 0)
617 return ret;
618
619 codec->width = ost->output_video_filter->inputs[0]->w;
620 codec->height = ost->output_video_filter->inputs[0]->h;
621 codec->sample_aspect_ratio = ost->st->sample_aspect_ratio =
622 ost->frame_aspect_ratio ? // overridden by the -aspect cli option
623 av_d2q(ost->frame_aspect_ratio * codec->height/codec->width, 255) :
624 ost->output_video_filter->inputs[0]->sample_aspect_ratio;
625
626 return 0;
627 }
628 #endif /* CONFIG_AVFILTER */
629
630 static void term_exit(void)
631 {
632 av_log(NULL, AV_LOG_QUIET, "");
633 }
634
635 static volatile int received_sigterm = 0;
636 static volatile int received_nb_signals = 0;
637
638 static void
639 sigterm_handler(int sig)
640 {
641 received_sigterm = sig;
642 received_nb_signals++;
643 term_exit();
644 }
645
646 static void term_init(void)
647 {
648 signal(SIGINT , sigterm_handler); /* Interrupt (ANSI). */
649 signal(SIGTERM, sigterm_handler); /* Termination (ANSI). */
650 #ifdef SIGXCPU
651 signal(SIGXCPU, sigterm_handler);
652 #endif
653 }
654
655 static int decode_interrupt_cb(void *ctx)
656 {
657 return received_nb_signals > 1;
658 }
659
660 static const AVIOInterruptCB int_cb = { decode_interrupt_cb, NULL };
661
662 void exit_program(int ret)
663 {
664 int i;
665
666 /* close files */
667 for (i = 0; i < nb_output_files; i++) {
668 AVFormatContext *s = output_files[i].ctx;
669 if (!(s->oformat->flags & AVFMT_NOFILE) && s->pb)
670 avio_close(s->pb);
671 avformat_free_context(s);
672 av_dict_free(&output_files[i].opts);
673 }
674 for (i = 0; i < nb_output_streams; i++) {
675 AVBitStreamFilterContext *bsfc = output_streams[i].bitstream_filters;
676 while (bsfc) {
677 AVBitStreamFilterContext *next = bsfc->next;
678 av_bitstream_filter_close(bsfc);
679 bsfc = next;
680 }
681 output_streams[i].bitstream_filters = NULL;
682
683 if (output_streams[i].output_frame) {
684 AVFrame *frame = output_streams[i].output_frame;
685 if (frame->extended_data != frame->data)
686 av_freep(&frame->extended_data);
687 av_freep(&frame);
688 }
689
690 #if CONFIG_AVFILTER
691 av_freep(&output_streams[i].avfilter);
692 #endif
693 }
694 for (i = 0; i < nb_input_files; i++) {
695 avformat_close_input(&input_files[i].ctx);
696 }
697 for (i = 0; i < nb_input_streams; i++) {
698 av_freep(&input_streams[i].decoded_frame);
699 av_freep(&input_streams[i].filtered_frame);
700 av_dict_free(&input_streams[i].opts);
701 free_buffer_pool(&input_streams[i]);
702 }
703
704 if (vstats_file)
705 fclose(vstats_file);
706 av_free(vstats_filename);
707
708 av_freep(&input_streams);
709 av_freep(&input_files);
710 av_freep(&output_streams);
711 av_freep(&output_files);
712
713 uninit_opts();
714 av_free(audio_buf);
715 allocated_audio_buf_size = 0;
716
717 #if CONFIG_AVFILTER
718 avfilter_uninit();
719 #endif
720 avformat_network_deinit();
721
722 if (received_sigterm) {
723 av_log(NULL, AV_LOG_INFO, "Received signal %d: terminating.\n",
724 (int) received_sigterm);
725 exit (255);
726 }
727
728 exit(ret);
729 }
730
731 static void assert_avoptions(AVDictionary *m)
732 {
733 AVDictionaryEntry *t;
734 if ((t = av_dict_get(m, "", NULL, AV_DICT_IGNORE_SUFFIX))) {
735 av_log(NULL, AV_LOG_FATAL, "Option %s not found.\n", t->key);
736 exit_program(1);
737 }
738 }
739
740 static void assert_codec_experimental(AVCodecContext *c, int encoder)
741 {
742 const char *codec_string = encoder ? "encoder" : "decoder";
743 AVCodec *codec;
744 if (c->codec->capabilities & CODEC_CAP_EXPERIMENTAL &&
745 c->strict_std_compliance > FF_COMPLIANCE_EXPERIMENTAL) {
746 av_log(NULL, AV_LOG_FATAL, "%s '%s' is experimental and might produce bad "
747 "results.\nAdd '-strict experimental' if you want to use it.\n",
748 codec_string, c->codec->name);
749 codec = encoder ? avcodec_find_encoder(c->codec->id) : avcodec_find_decoder(c->codec->id);
750 if (!(codec->capabilities & CODEC_CAP_EXPERIMENTAL))
751 av_log(NULL, AV_LOG_FATAL, "Or use the non experimental %s '%s'.\n",
752 codec_string, codec->name);
753 exit_program(1);
754 }
755 }
756
757 static void choose_sample_fmt(AVStream *st, AVCodec *codec)
758 {
759 if (codec && codec->sample_fmts) {
760 const enum AVSampleFormat *p = codec->sample_fmts;
761 for (; *p != -1; p++) {
762 if (*p == st->codec->sample_fmt)
763 break;
764 }
765 if (*p == -1) {
766 av_log(NULL, AV_LOG_WARNING,
767 "Incompatible sample format '%s' for codec '%s', auto-selecting format '%s'\n",
768 av_get_sample_fmt_name(st->codec->sample_fmt),
769 codec->name,
770 av_get_sample_fmt_name(codec->sample_fmts[0]));
771 st->codec->sample_fmt = codec->sample_fmts[0];
772 }
773 }
774 }
775
776 /**
777 * Update the requested input sample format based on the output sample format.
778 * This is currently only used to request float output from decoders which
779 * support multiple sample formats, one of which is AV_SAMPLE_FMT_FLT.
780 * Ideally this will be removed in the future when decoders do not do format
781 * conversion and only output in their native format.
782 */
783 static void update_sample_fmt(AVCodecContext *dec, AVCodec *dec_codec,
784 AVCodecContext *enc)
785 {
786 /* if sample formats match or a decoder sample format has already been
787 requested, just return */
788 if (enc->sample_fmt == dec->sample_fmt ||
789 dec->request_sample_fmt > AV_SAMPLE_FMT_NONE)
790 return;
791
792 /* if decoder supports more than one output format */
793 if (dec_codec && dec_codec->sample_fmts &&
794 dec_codec->sample_fmts[0] != AV_SAMPLE_FMT_NONE &&
795 dec_codec->sample_fmts[1] != AV_SAMPLE_FMT_NONE) {
796 const enum AVSampleFormat *p;
797 int min_dec = -1, min_inc = -1;
798
799 /* find a matching sample format in the encoder */
800 for (p = dec_codec->sample_fmts; *p != AV_SAMPLE_FMT_NONE; p++) {
801 if (*p == enc->sample_fmt) {
802 dec->request_sample_fmt = *p;
803 return;
804 } else if (*p > enc->sample_fmt) {
805 min_inc = FFMIN(min_inc, *p - enc->sample_fmt);
806 } else
807 min_dec = FFMIN(min_dec, enc->sample_fmt - *p);
808 }
809
810 /* if none match, provide the one that matches quality closest */
811 dec->request_sample_fmt = min_inc > 0 ? enc->sample_fmt + min_inc :
812 enc->sample_fmt - min_dec;
813 }
814 }
815
816 static void choose_sample_rate(AVStream *st, AVCodec *codec)
817 {
818 if (codec && codec->supported_samplerates) {
819 const int *p = codec->supported_samplerates;
820 int best = 0;
821 int best_dist = INT_MAX;
822 for (; *p; p++) {
823 int dist = abs(st->codec->sample_rate - *p);
824 if (dist < best_dist) {
825 best_dist = dist;
826 best = *p;
827 }
828 }
829 if (best_dist) {
830 av_log(st->codec, AV_LOG_WARNING, "Requested sampling rate unsupported using closest supported (%d)\n", best);
831 }
832 st->codec->sample_rate = best;
833 }
834 }
835
836 static void choose_pixel_fmt(AVStream *st, AVCodec *codec)
837 {
838 if (codec && codec->pix_fmts) {
839 const enum PixelFormat *p = codec->pix_fmts;
840 if (st->codec->strict_std_compliance <= FF_COMPLIANCE_UNOFFICIAL) {
841 if (st->codec->codec_id == CODEC_ID_MJPEG) {
842 p = (const enum PixelFormat[]) { PIX_FMT_YUVJ420P, PIX_FMT_YUVJ422P, PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_NONE };
843 } else if (st->codec->codec_id == CODEC_ID_LJPEG) {
844 p = (const enum PixelFormat[]) { PIX_FMT_YUVJ420P, PIX_FMT_YUVJ422P, PIX_FMT_YUVJ444P, PIX_FMT_YUV420P,
845 PIX_FMT_YUV422P, PIX_FMT_YUV444P, PIX_FMT_BGRA, PIX_FMT_NONE };
846 }
847 }
848 for (; *p != PIX_FMT_NONE; p++) {
849 if (*p == st->codec->pix_fmt)
850 break;
851 }
852 if (*p == PIX_FMT_NONE) {
853 if (st->codec->pix_fmt != PIX_FMT_NONE)
854 av_log(NULL, AV_LOG_WARNING,
855 "Incompatible pixel format '%s' for codec '%s', auto-selecting format '%s'\n",
856 av_pix_fmt_descriptors[st->codec->pix_fmt].name,
857 codec->name,
858 av_pix_fmt_descriptors[codec->pix_fmts[0]].name);
859 st->codec->pix_fmt = codec->pix_fmts[0];
860 }
861 }
862 }
863
864 static double
865 get_sync_ipts(const OutputStream *ost)
866 {
867 const InputStream *ist = ost->sync_ist;
868 OutputFile *of = &output_files[ost->file_index];
869 return (double)(ist->pts - of->start_time) / AV_TIME_BASE;
870 }
871
872 static void write_frame(AVFormatContext *s, AVPacket *pkt, OutputStream *ost)
873 {
874 AVBitStreamFilterContext *bsfc = ost->bitstream_filters;
875 AVCodecContext *avctx = ost->st->codec;
876 int ret;
877
878 while (bsfc) {
879 AVPacket new_pkt = *pkt;
880 int a = av_bitstream_filter_filter(bsfc, avctx, NULL,
881 &new_pkt.data, &new_pkt.size,
882 pkt->data, pkt->size,
883 pkt->flags & AV_PKT_FLAG_KEY);
884 if (a > 0) {
885 av_free_packet(pkt);
886 new_pkt.destruct = av_destruct_packet;
887 } else if (a < 0) {
888 av_log(NULL, AV_LOG_ERROR, "%s failed for stream %d, codec %s",
889 bsfc->filter->name, pkt->stream_index,
890 avctx->codec ? avctx->codec->name : "copy");
891 print_error("", a);
892 if (exit_on_error)
893 exit_program(1);
894 }
895 *pkt = new_pkt;
896
897 bsfc = bsfc->next;
898 }
899
900 ret = av_interleaved_write_frame(s, pkt);
901 if (ret < 0) {
902 print_error("av_interleaved_write_frame()", ret);
903 exit_program(1);
904 }
905 ost->frame_number++;
906 }
907
908 static void generate_silence(uint8_t* buf, enum AVSampleFormat sample_fmt, size_t size)
909 {
910 int fill_char = 0x00;
911 if (sample_fmt == AV_SAMPLE_FMT_U8)
912 fill_char = 0x80;
913 memset(buf, fill_char, size);
914 }
915
916 static int encode_audio_frame(AVFormatContext *s, OutputStream *ost,
917 const uint8_t *buf, int buf_size)
918 {
919 AVCodecContext *enc = ost->st->codec;
920 AVFrame *frame = NULL;
921 AVPacket pkt;
922 int ret, got_packet;
923
924 av_init_packet(&pkt);
925 pkt.data = NULL;
926 pkt.size = 0;
927
928 if (buf) {
929 if (!ost->output_frame) {
930 ost->output_frame = avcodec_alloc_frame();
931 if (!ost->output_frame) {
932 av_log(NULL, AV_LOG_FATAL, "out-of-memory in encode_audio_frame()\n");
933 exit_program(1);
934 }
935 }
936 frame = ost->output_frame;
937 if (frame->extended_data != frame->data)
938 av_freep(&frame->extended_data);
939 avcodec_get_frame_defaults(frame);
940
941 frame->nb_samples = buf_size /
942 (enc->channels * av_get_bytes_per_sample(enc->sample_fmt));
943 if ((ret = avcodec_fill_audio_frame(frame, enc->channels, enc->sample_fmt,
944 buf, buf_size, 1)) < 0) {
945 av_log(NULL, AV_LOG_FATAL, "Audio encoding failed\n");
946 exit_program(1);
947 }
948 }
949
950 got_packet = 0;
951 if (avcodec_encode_audio2(enc, &pkt, frame, &got_packet) < 0) {
952 av_log(NULL, AV_LOG_FATAL, "Audio encoding failed\n");
953 exit_program(1);
954 }
955
956 if (got_packet) {
957 pkt.stream_index = ost->index;
958 if (pkt.pts != AV_NOPTS_VALUE)
959 pkt.pts = av_rescale_q(pkt.pts, enc->time_base, ost->st->time_base);
960 if (pkt.duration > 0)
961 pkt.duration = av_rescale_q(pkt.duration, enc->time_base, ost->st->time_base);
962
963 write_frame(s, &pkt, ost);
964
965 audio_size += pkt.size;
966 }
967
968 if (frame)
969 ost->sync_opts += frame->nb_samples;
970
971 return pkt.size;
972 }
973
974 static void do_audio_out(AVFormatContext *s, OutputStream *ost,
975 InputStream *ist, AVFrame *decoded_frame)
976 {
977 uint8_t *buftmp;
978 int64_t audio_buf_size;
979
980 int size_out, frame_bytes, resample_changed;
981 AVCodecContext *enc = ost->st->codec;
982 AVCodecContext *dec = ist->st->codec;
983 int osize = av_get_bytes_per_sample(enc->sample_fmt);
984 int isize = av_get_bytes_per_sample(dec->sample_fmt);
985 uint8_t *buf = decoded_frame->data[0];
986 int size = decoded_frame->nb_samples * dec->channels * isize;
987 int64_t allocated_for_size = size;
988
989 need_realloc:
990 audio_buf_size = (allocated_for_size + isize * dec->channels - 1) / (isize * dec->channels);
991 audio_buf_size = (audio_buf_size * enc->sample_rate + dec->sample_rate) / dec->sample_rate;
992 audio_buf_size = audio_buf_size * 2 + 10000; // safety factors for the deprecated resampling API
993 audio_buf_size = FFMAX(audio_buf_size, enc->frame_size);
994 audio_buf_size *= osize * enc->channels;
995
996 if (audio_buf_size > INT_MAX) {
997 av_log(NULL, AV_LOG_FATAL, "Buffer sizes too large\n");
998 exit_program(1);
999 }
1000
1001 av_fast_malloc(&audio_buf, &allocated_audio_buf_size, audio_buf_size);
1002 if (!audio_buf) {
1003 av_log(NULL, AV_LOG_FATAL, "Out of memory in do_audio_out\n");
1004 exit_program(1);
1005 }
1006
1007 if (enc->channels != dec->channels || enc->sample_rate != dec->sample_rate)
1008 ost->audio_resample = 1;
1009
1010 resample_changed = ost->resample_sample_fmt != dec->sample_fmt ||
1011 ost->resample_channels != dec->channels ||
1012 ost->resample_sample_rate != dec->sample_rate;
1013
1014 if ((ost->audio_resample && !ost->resample) || resample_changed) {
1015 if (resample_changed) {
1016 av_log(NULL, AV_LOG_INFO, "Input stream #%d:%d frame changed from rate:%d fmt:%s ch:%d to rate:%d fmt:%s ch:%d\n",
1017 ist->file_index, ist->st->index,
1018 ost->resample_sample_rate, av_get_sample_fmt_name(ost->resample_sample_fmt), ost->resample_channels,
1019 dec->sample_rate, av_get_sample_fmt_name(dec->sample_fmt), dec->channels);
1020 ost->resample_sample_fmt = dec->sample_fmt;
1021 ost->resample_channels = dec->channels;
1022 ost->resample_sample_rate = dec->sample_rate;
1023 if (ost->resample)
1024 audio_resample_close(ost->resample);
1025 }
1026 /* if audio_sync_method is >1 the resampler is needed for audio drift compensation */
1027 if (audio_sync_method <= 1 &&
1028 ost->resample_sample_fmt == enc->sample_fmt &&
1029 ost->resample_channels == enc->channels &&
1030 ost->resample_sample_rate == enc->sample_rate) {
1031 ost->resample = NULL;
1032 ost->audio_resample = 0;
1033 } else if (ost->audio_resample) {
1034 if (dec->sample_fmt != AV_SAMPLE_FMT_S16)
1035 av_log(NULL, AV_LOG_WARNING, "Using s16 intermediate sample format for resampling\n");
1036 ost->resample = av_audio_resample_init(enc->channels, dec->channels,
1037 enc->sample_rate, dec->sample_rate,
1038 enc->sample_fmt, dec->sample_fmt,
1039 16, 10, 0, 0.8);
1040 if (!ost->resample) {
1041 av_log(NULL, AV_LOG_FATAL, "Can not resample %d channels @ %d Hz to %d channels @ %d Hz\n",
1042 dec->channels, dec->sample_rate,
1043 enc->channels, enc->sample_rate);
1044 exit_program(1);
1045 }
1046 }
1047 }
1048
1049 #define MAKE_SFMT_PAIR(a,b) ((a)+AV_SAMPLE_FMT_NB*(b))
1050 if (!ost->audio_resample && dec->sample_fmt != enc->sample_fmt &&
1051 MAKE_SFMT_PAIR(enc->sample_fmt,dec->sample_fmt) != ost->reformat_pair) {
1052 if (ost->reformat_ctx)
1053 av_audio_convert_free(ost->reformat_ctx);
1054 ost->reformat_ctx = av_audio_convert_alloc(enc->sample_fmt, 1,
1055 dec->sample_fmt, 1, NULL, 0);
1056 if (!ost->reformat_ctx) {
1057 av_log(NULL, AV_LOG_FATAL, "Cannot convert %s sample format to %s sample format\n",
1058 av_get_sample_fmt_name(dec->sample_fmt),
1059 av_get_sample_fmt_name(enc->sample_fmt));
1060 exit_program(1);
1061 }
1062 ost->reformat_pair = MAKE_SFMT_PAIR(enc->sample_fmt,dec->sample_fmt);
1063 }
1064
1065 if (audio_sync_method) {
1066 double delta = get_sync_ipts(ost) * enc->sample_rate - ost->sync_opts -
1067 av_fifo_size(ost->fifo) / (enc->channels * osize);
1068 int idelta = delta * dec->sample_rate / enc->sample_rate;
1069 int byte_delta = idelta * isize * dec->channels;
1070
1071 // FIXME resample delay
1072 if (fabs(delta) > 50) {
1073 if (ist->is_start || fabs(delta) > audio_drift_threshold*enc->sample_rate) {
1074 if (byte_delta < 0) {
1075 byte_delta = FFMAX(byte_delta, -size);
1076 size += byte_delta;
1077 buf -= byte_delta;
1078 av_log(NULL, AV_LOG_VERBOSE, "discarding %d audio samples\n",
1079 -byte_delta / (isize * dec->channels));
1080 if (!size)
1081 return;
1082 ist->is_start = 0;
1083 } else {
1084 static uint8_t *input_tmp = NULL;
1085 input_tmp = av_realloc(input_tmp, byte_delta + size);
1086
1087 if (byte_delta > allocated_for_size - size) {
1088 allocated_for_size = byte_delta + (int64_t)size;
1089 goto need_realloc;
1090 }
1091 ist->is_start = 0;
1092
1093 generate_silence(input_tmp, dec->sample_fmt, byte_delta);
1094 memcpy(input_tmp + byte_delta, buf, size);
1095 buf = input_tmp;
1096 size += byte_delta;
1097 av_log(NULL, AV_LOG_VERBOSE, "adding %d audio samples of silence\n", idelta);
1098 }
1099 } else if (audio_sync_method > 1) {
1100 int comp = av_clip(delta, -audio_sync_method, audio_sync_method);
1101 av_assert0(ost->audio_resample);
1102 av_log(NULL, AV_LOG_VERBOSE, "compensating audio timestamp drift:%f compensation:%d in:%d\n",
1103 delta, comp, enc->sample_rate);
1104 // fprintf(stderr, "drift:%f len:%d opts:%"PRId64" ipts:%"PRId64" fifo:%d\n", delta, -1, ost->sync_opts, (int64_t)(get_sync_ipts(ost) * enc->sample_rate), av_fifo_size(ost->fifo)/(ost->st->codec->channels * 2));
1105 av_resample_compensate(*(struct AVResampleContext**)ost->resample, comp, enc->sample_rate);
1106 }
1107 }
1108 } else
1109 ost->sync_opts = lrintf(get_sync_ipts(ost) * enc->sample_rate) -
1110 av_fifo_size(ost->fifo) / (enc->channels * osize); // FIXME wrong
1111
1112 if (ost->audio_resample) {
1113 buftmp = audio_buf;
1114 size_out = audio_resample(ost->resample,
1115 (short *)buftmp, (short *)buf,
1116 size / (dec->channels * isize));
1117 size_out = size_out * enc->channels * osize;
1118 } else {
1119 buftmp = buf;
1120 size_out = size;
1121 }
1122
1123 if (!ost->audio_resample && dec->sample_fmt != enc->sample_fmt) {
1124 const void *ibuf[6] = { buftmp };
1125 void *obuf[6] = { audio_buf };
1126 int istride[6] = { isize };
1127 int ostride[6] = { osize };
1128 int len = size_out / istride[0];
1129 if (av_audio_convert(ost->reformat_ctx, obuf, ostride, ibuf, istride, len) < 0) {
1130 printf("av_audio_convert() failed\n");
1131 if (exit_on_error)
1132 exit_program(1);
1133 return;
1134 }
1135 buftmp = audio_buf;
1136 size_out = len * osize;
1137 }
1138
1139 /* now encode as many frames as possible */
1140 if (!(enc->codec->capabilities & CODEC_CAP_VARIABLE_FRAME_SIZE)) {
1141 /* output resampled raw samples */
1142 if (av_fifo_realloc2(ost->fifo, av_fifo_size(ost->fifo) + size_out) < 0) {
1143 av_log(NULL, AV_LOG_FATAL, "av_fifo_realloc2() failed\n");
1144 exit_program(1);
1145 }
1146 av_fifo_generic_write(ost->fifo, buftmp, size_out, NULL);
1147
1148 frame_bytes = enc->frame_size * osize * enc->channels;
1149
1150 while (av_fifo_size(ost->fifo) >= frame_bytes) {
1151 av_fifo_generic_read(ost->fifo, audio_buf, frame_bytes, NULL);
1152 encode_audio_frame(s, ost, audio_buf, frame_bytes);
1153 }
1154 } else {
1155 encode_audio_frame(s, ost, buftmp, size_out);
1156 }
1157 }
1158
1159 static void pre_process_video_frame(InputStream *ist, AVPicture *picture, void **bufp)
1160 {
1161 AVCodecContext *dec;
1162 AVPicture *picture2;
1163 AVPicture picture_tmp;
1164 uint8_t *buf = 0;
1165
1166 dec = ist->st->codec;
1167
1168 /* deinterlace : must be done before any resize */
1169 if (do_deinterlace) {
1170 int size;
1171
1172 /* create temporary picture */
1173 size = avpicture_get_size(dec->pix_fmt, dec->width, dec->height);
1174 buf = av_malloc(size);
1175 if (!buf)
1176 return;
1177
1178 picture2 = &picture_tmp;
1179 avpicture_fill(picture2, buf, dec->pix_fmt, dec->width, dec->height);
1180
1181 if (avpicture_deinterlace(picture2, picture,
1182 dec->pix_fmt, dec->width, dec->height) < 0) {
1183 /* if error, do not deinterlace */
1184 av_log(NULL, AV_LOG_WARNING, "Deinterlacing failed\n");
1185 av_free(buf);
1186 buf = NULL;
1187 picture2 = picture;
1188 }
1189 } else {
1190 picture2 = picture;
1191 }
1192
1193 if (picture != picture2)
1194 *picture = *picture2;
1195 *bufp = buf;
1196 }
1197
1198 static void do_subtitle_out(AVFormatContext *s,
1199 OutputStream *ost,
1200 InputStream *ist,
1201 AVSubtitle *sub,
1202 int64_t pts)
1203 {
1204 static uint8_t *subtitle_out = NULL;
1205 int subtitle_out_max_size = 1024 * 1024;
1206 int subtitle_out_size, nb, i;
1207 AVCodecContext *enc;
1208 AVPacket pkt;
1209
1210 if (pts == AV_NOPTS_VALUE) {
1211 av_log(NULL, AV_LOG_ERROR, "Subtitle packets must have a pts\n");
1212 if (exit_on_error)
1213 exit_program(1);
1214 return;
1215 }
1216
1217 enc = ost->st->codec;
1218
1219 if (!subtitle_out) {
1220 subtitle_out = av_malloc(subtitle_out_max_size);
1221 }
1222
1223 /* Note: DVB subtitle need one packet to draw them and one other
1224 packet to clear them */
1225 /* XXX: signal it in the codec context ? */
1226 if (enc->codec_id == CODEC_ID_DVB_SUBTITLE)
1227 nb = 2;
1228 else
1229 nb = 1;
1230
1231 for (i = 0; i < nb; i++) {
1232 sub->pts = av_rescale_q(pts, ist->st->time_base, AV_TIME_BASE_Q);
1233 // start_display_time is required to be 0
1234 sub->pts += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, AV_TIME_BASE_Q);
1235 sub->end_display_time -= sub->start_display_time;
1236 sub->start_display_time = 0;
1237 subtitle_out_size = avcodec_encode_subtitle(enc, subtitle_out,
1238 subtitle_out_max_size, sub);
1239 if (subtitle_out_size < 0) {
1240 av_log(NULL, AV_LOG_FATAL, "Subtitle encoding failed\n");
1241 exit_program(1);
1242 }
1243
1244 av_init_packet(&pkt);
1245 pkt.stream_index = ost->index;
1246 pkt.data = subtitle_out;
1247 pkt.size = subtitle_out_size;
1248 pkt.pts = av_rescale_q(sub->pts, AV_TIME_BASE_Q, ost->st->time_base);
1249 if (enc->codec_id == CODEC_ID_DVB_SUBTITLE) {
1250 /* XXX: the pts correction is handled here. Maybe handling
1251 it in the codec would be better */
1252 if (i == 0)
1253 pkt.pts += 90 * sub->start_display_time;
1254 else
1255 pkt.pts += 90 * sub->end_display_time;
1256 }
1257 write_frame(s, &pkt, ost);
1258 }
1259 }
1260
1261 static int bit_buffer_size = 1024 * 256;
1262 static uint8_t *bit_buffer = NULL;
1263
1264 #if !CONFIG_AVFILTER
1265 static void do_video_resample(OutputStream *ost,
1266 InputStream *ist,
1267 AVFrame *in_picture,
1268 AVFrame **out_picture)
1269 {
1270 int resample_changed = 0;
1271 *out_picture = in_picture;
1272
1273 resample_changed = ost->resample_width != in_picture->width ||
1274 ost->resample_height != in_picture->height ||
1275 ost->resample_pix_fmt != in_picture->format;
1276
1277 if (resample_changed) {
1278 av_log(NULL, AV_LOG_INFO,
1279 "Input stream #%d:%d frame changed from size:%dx%d fmt:%s to size:%dx%d fmt:%s\n",
1280 ist->file_index, ist->st->index,
1281 ost->resample_width, ost->resample_height, av_get_pix_fmt_name(ost->resample_pix_fmt),
1282 in_picture->width, in_picture->height, av_get_pix_fmt_name(in_picture->format));
1283 if (!ost->video_resample)
1284 ost->video_resample = 1;
1285 }
1286
1287 if (ost->video_resample) {
1288 *out_picture = &ost->pict_tmp;
1289 if (resample_changed) {
1290 /* initialize a new scaler context */
1291 sws_freeContext(ost->img_resample_ctx);
1292 ost->img_resample_ctx = sws_getContext(
1293 ist->st->codec->width,
1294 ist->st->codec->height,
1295 ist->st->codec->pix_fmt,
1296 ost->st->codec->width,
1297 ost->st->codec->height,
1298 ost->st->codec->pix_fmt,
1299 ost->sws_flags, NULL, NULL, NULL);
1300 if (ost->img_resample_ctx == NULL) {
1301 av_log(NULL, AV_LOG_FATAL, "Cannot get resampling context\n");
1302 exit_program(1);
1303 }
1304 }
1305 sws_scale(ost->img_resample_ctx, in_picture->data, in_picture->linesize,
1306 0, ost->resample_height, (*out_picture)->data, (*out_picture)->linesize);
1307 }
1308 if (resample_changed) {
1309 ost->resample_width = in_picture->width;
1310 ost->resample_height = in_picture->height;
1311 ost->resample_pix_fmt = in_picture->format;
1312 }
1313 }
1314 #endif
1315
1316
1317 static void do_video_out(AVFormatContext *s,
1318 OutputStream *ost,
1319 InputStream *ist,
1320 AVFrame *in_picture,
1321 int *frame_size, float quality)
1322 {
1323 int nb_frames, i, ret, format_video_sync;
1324 AVFrame *final_picture;
1325 AVCodecContext *enc;
1326 double sync_ipts;
1327
1328 enc = ost->st->codec;
1329
1330 sync_ipts = get_sync_ipts(ost) / av_q2d(enc->time_base);
1331
1332 /* by default, we output a single frame */
1333 nb_frames = 1;
1334
1335 *frame_size = 0;
1336
1337 format_video_sync = video_sync_method;
1338 if (format_video_sync == VSYNC_AUTO)
1339 format_video_sync = (s->oformat->flags & AVFMT_NOTIMESTAMPS) ? VSYNC_PASSTHROUGH :
1340 (s->oformat->flags & AVFMT_VARIABLE_FPS) ? VSYNC_VFR : VSYNC_CFR;
1341
1342 if (format_video_sync != VSYNC_PASSTHROUGH) {
1343 double vdelta = sync_ipts - ost->sync_opts;
1344 // FIXME set to 0.5 after we fix some dts/pts bugs like in avidec.c
1345 if (vdelta < -1.1)
1346 nb_frames = 0;
1347 else if (format_video_sync == VSYNC_VFR) {
1348 if (vdelta <= -0.6) {
1349 nb_frames = 0;
1350 } else if (vdelta > 0.6)
1351 ost->sync_opts = lrintf(sync_ipts);
1352 } else if (vdelta > 1.1)
1353 nb_frames = lrintf(vdelta);
1354 //fprintf(stderr, "vdelta:%f, ost->sync_opts:%"PRId64", ost->sync_ipts:%f nb_frames:%d\n", vdelta, ost->sync_opts, get_sync_ipts(ost), nb_frames);
1355 if (nb_frames == 0) {
1356 ++nb_frames_drop;
1357 av_log(NULL, AV_LOG_VERBOSE, "*** drop!\n");
1358 } else if (nb_frames > 1) {
1359 nb_frames_dup += nb_frames - 1;
1360 av_log(NULL, AV_LOG_VERBOSE, "*** %d dup!\n", nb_frames - 1);
1361 }
1362 } else
1363 ost->sync_opts = lrintf(sync_ipts);
1364
1365 nb_frames = FFMIN(nb_frames, ost->max_frames - ost->frame_number);
1366 if (nb_frames <= 0)
1367 return;
1368
1369 #if !CONFIG_AVFILTER
1370 do_video_resample(ost, ist, in_picture, &final_picture);
1371 #else
1372 final_picture = in_picture;
1373 #endif
1374
1375 /* duplicates frame if needed */
1376 for (i = 0; i < nb_frames; i++) {
1377 AVPacket pkt;
1378 av_init_packet(&pkt);
1379 pkt.stream_index = ost->index;
1380
1381 if (s->oformat->flags & AVFMT_RAWPICTURE &&
1382 enc->codec->id == CODEC_ID_RAWVIDEO) {
1383 /* raw pictures are written as AVPicture structure to
1384 avoid any copies. We support temporarily the older
1385 method. */
1386 enc->coded_frame->interlaced_frame = in_picture->interlaced_frame;
1387 enc->coded_frame->top_field_first = in_picture->top_field_first;
1388 pkt.data = (uint8_t *)final_picture;
1389 pkt.size = sizeof(AVPicture);
1390 pkt.pts = av_rescale_q(ost->sync_opts, enc->time_base, ost->st->time_base);
1391 pkt.flags |= AV_PKT_FLAG_KEY;
1392
1393 write_frame(s, &pkt, ost);
1394 } else {
1395 AVFrame big_picture;
1396
1397 big_picture = *final_picture;
1398 /* better than nothing: use input picture interlaced
1399 settings */
1400 big_picture.interlaced_frame = in_picture->interlaced_frame;
1401 if (ost->st->codec->flags & (CODEC_FLAG_INTERLACED_DCT|CODEC_FLAG_INTERLACED_ME)) {
1402 if (ost->top_field_first == -1)
1403 big_picture.top_field_first = in_picture->top_field_first;
1404 else
1405 big_picture.top_field_first = !!ost->top_field_first;
1406 }
1407
1408 /* handles same_quant here. This is not correct because it may
1409 not be a global option */
1410 big_picture.quality = quality;
1411 if (!enc->me_threshold)
1412 big_picture.pict_type = 0;
1413 // big_picture.pts = AV_NOPTS_VALUE;
1414 big_picture.pts = ost->sync_opts;
1415 // big_picture.pts= av_rescale(ost->sync_opts, AV_TIME_BASE*(int64_t)enc->time_base.num, enc->time_base.den);
1416 // av_log(NULL, AV_LOG_DEBUG, "%"PRId64" -> encoder\n", ost->sync_opts);
1417 if (ost->forced_kf_index < ost->forced_kf_count &&
1418 big_picture.pts >= ost->forced_kf_pts[ost->forced_kf_index]) {
1419 big_picture.pict_type = AV_PICTURE_TYPE_I;
1420 ost->forced_kf_index++;
1421 }
1422 ret = avcodec_encode_video(enc,
1423 bit_buffer, bit_buffer_size,
1424 &big_picture);
1425 if (ret < 0) {
1426 av_log(NULL, AV_LOG_FATAL, "Video encoding failed\n");
1427 exit_program(1);
1428 }
1429
1430 if (ret > 0) {
1431 pkt.data = bit_buffer;
1432 pkt.size = ret;
1433 if (enc->coded_frame->pts != AV_NOPTS_VALUE)
1434 pkt.pts = av_rescale_q(enc->coded_frame->pts, enc->time_base, ost->st->time_base);
1435 /*av_log(NULL, AV_LOG_DEBUG, "encoder -> %"PRId64"/%"PRId64"\n",
1436 pkt.pts != AV_NOPTS_VALUE ? av_rescale(pkt.pts, enc->time_base.den, AV_TIME_BASE*(int64_t)enc->time_base.num) : -1,
1437 pkt.dts != AV_NOPTS_VALUE ? av_rescale(pkt.dts, enc->time_base.den, AV_TIME_BASE*(int64_t)enc->time_base.num) : -1);*/
1438
1439 if (enc->coded_frame->key_frame)
1440 pkt.flags |= AV_PKT_FLAG_KEY;
1441 write_frame(s, &pkt, ost);
1442 *frame_size = ret;
1443 video_size += ret;
1444 // fprintf(stderr,"\nFrame: %3d size: %5d type: %d",
1445 // enc->frame_number-1, ret, enc->pict_type);
1446 /* if two pass, output log */
1447 if (ost->logfile && enc->stats_out) {
1448 fprintf(ost->logfile, "%s", enc->stats_out);
1449 }
1450 }
1451 }
1452 ost->sync_opts++;
1453 }
1454 }
1455
1456 static double psnr(double d)
1457 {
1458 return -10.0 * log(d) / log(10.0);
1459 }
1460
1461 static void do_video_stats(AVFormatContext *os, OutputStream *ost,
1462 int frame_size)
1463 {
1464 AVCodecContext *enc;
1465 int frame_number;
1466 double ti1, bitrate, avg_bitrate;
1467
1468 /* this is executed just the first time do_video_stats is called */
1469 if (!vstats_file) {
1470 vstats_file = fopen(vstats_filename, "w");
1471 if (!vstats_file) {
1472 perror("fopen");
1473 exit_program(1);
1474 }
1475 }
1476
1477 enc = ost->st->codec;
1478 if (enc->codec_type == AVMEDIA_TYPE_VIDEO) {
1479 frame_number = ost->frame_number;
1480 fprintf(vstats_file, "frame= %5d q= %2.1f ", frame_number, enc->coded_frame->quality / (float)FF_QP2LAMBDA);
1481 if (enc->flags&CODEC_FLAG_PSNR)
1482 fprintf(vstats_file, "PSNR= %6.2f ", psnr(enc->coded_frame->error[0] / (enc->width * enc->height * 255.0 * 255.0)));
1483
1484 fprintf(vstats_file,"f_size= %6d ", frame_size);
1485 /* compute pts value */
1486 ti1 = ost->sync_opts * av_q2d(enc->time_base);
1487 if (ti1 < 0.01)
1488 ti1 = 0.01;
1489
1490 bitrate = (frame_size * 8) / av_q2d(enc->time_base) / 1000.0;
1491 avg_bitrate = (double)(video_size * 8) / ti1 / 1000.0;
1492 fprintf(vstats_file, "s_size= %8.0fkB time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ",
1493 (double)video_size / 1024, ti1, bitrate, avg_bitrate);
1494 fprintf(vstats_file, "type= %c\n", av_get_picture_type_char(enc->coded_frame->pict_type));
1495 }
1496 }
1497
1498 static void print_report(OutputFile *output_files,
1499 OutputStream *ost_table, int nb_ostreams,
1500 int is_last_report, int64_t timer_start)
1501 {
1502 char buf[1024];
1503 OutputStream *ost;
1504 AVFormatContext *oc;
1505 int64_t total_size;
1506 AVCodecContext *enc;
1507 int frame_number, vid, i;
1508 double bitrate, ti1, pts;
1509 static int64_t last_time = -1;
1510 static int qp_histogram[52];
1511
1512 if (!print_stats && !is_last_report)
1513 return;
1514
1515 if (!is_last_report) {
1516 int64_t cur_time;
1517 /* display the report every 0.5 seconds */
1518 cur_time = av_gettime();
1519 if (last_time == -1) {
1520 last_time = cur_time;
1521 return;
1522 }
1523 if ((cur_time - last_time) < 500000)
1524 return;
1525 last_time = cur_time;
1526 }
1527
1528
1529 oc = output_files[0].ctx;
1530
1531 total_size = avio_size(oc->pb);
1532 if (total_size < 0) // FIXME improve avio_size() so it works with non seekable output too
1533 total_size = avio_tell(oc->pb);
1534
1535 buf[0] = '\0';
1536 ti1 = 1e10;
1537 vid = 0;
1538 for (i = 0; i < nb_ostreams; i++) {
1539 float q = -1;
1540 ost = &ost_table[i];
1541 enc = ost->st->codec;
1542 if (!ost->stream_copy && enc->coded_frame)
1543 q = enc->coded_frame->quality / (float)FF_QP2LAMBDA;
1544 if (vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {
1545 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "q=%2.1f ", q);
1546 }
1547 if (!vid && enc->codec_type == AVMEDIA_TYPE_VIDEO) {
1548 float t = (av_gettime() - timer_start) / 1000000.0;
1549
1550 frame_number = ost->frame_number;
1551 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "frame=%5d fps=%3d q=%3.1f ",
1552 frame_number, (t > 1) ? (int)(frame_number / t + 0.5) : 0, q);
1553 if (is_last_report)
1554 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "L");
1555 if (qp_hist) {
1556 int j;
1557 int qp = lrintf(q);
1558 if (qp >= 0 && qp < FF_ARRAY_ELEMS(qp_histogram))
1559 qp_histogram[qp]++;
1560 for (j = 0; j < 32; j++)
1561 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%X", (int)lrintf(log(qp_histogram[j] + 1) / log(2)));
1562 }
1563 if (enc->flags&CODEC_FLAG_PSNR) {
1564 int j;
1565 double error, error_sum = 0;
1566 double scale, scale_sum = 0;
1567 char type[3] = { 'Y','U','V' };
1568 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "PSNR=");
1569 for (j = 0; j < 3; j++) {
1570 if (is_last_report) {
1571 error = enc->error[j];
1572 scale = enc->width * enc->height * 255.0 * 255.0 * frame_number;
1573 } else {
1574 error = enc->coded_frame->error[j];
1575 scale = enc->width * enc->height * 255.0 * 255.0;
1576 }
1577 if (j)
1578 scale /= 4;
1579 error_sum += error;
1580 scale_sum += scale;
1581 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "%c:%2.2f ", type[j], psnr(error / scale));
1582 }
1583 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), "*:%2.2f ", psnr(error_sum / scale_sum));
1584 }
1585 vid = 1;
1586 }
1587 /* compute min output value */
1588 pts = (double)ost->st->pts.val * av_q2d(ost->st->time_base);
1589 if ((pts < ti1) && (pts > 0))
1590 ti1 = pts;
1591 }
1592 if (ti1 < 0.01)
1593 ti1 = 0.01;
1594
1595 bitrate = (double)(total_size * 8) / ti1 / 1000.0;
1596
1597 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf),
1598 "size=%8.0fkB time=%0.2f bitrate=%6.1fkbits/s",
1599 (double)total_size / 1024, ti1, bitrate);
1600
1601 if (nb_frames_dup || nb_frames_drop)
1602 snprintf(buf + strlen(buf), sizeof(buf) - strlen(buf), " dup=%d drop=%d",
1603 nb_frames_dup, nb_frames_drop);
1604
1605 av_log(NULL, AV_LOG_INFO, "%s \r", buf);
1606
1607 fflush(stderr);
1608
1609 if (is_last_report) {
1610 int64_t raw= audio_size + video_size + extra_size;
1611 av_log(NULL, AV_LOG_INFO, "\n");
1612 av_log(NULL, AV_LOG_INFO, "video:%1.0fkB audio:%1.0fkB global headers:%1.0fkB muxing overhead %f%%\n",
1613 video_size / 1024.0,
1614 audio_size / 1024.0,
1615 extra_size / 1024.0,
1616 100.0 * (total_size - raw) / raw
1617 );
1618 }
1619 }
1620
1621 static void flush_encoders(OutputStream *ost_table, int nb_ostreams)
1622 {
1623 int i, ret;
1624
1625 for (i = 0; i < nb_ostreams; i++) {
1626 OutputStream *ost = &ost_table[i];
1627 AVCodecContext *enc = ost->st->codec;
1628 AVFormatContext *os = output_files[ost->file_index].ctx;
1629 int stop_encoding = 0;
1630
1631 if (!ost->encoding_needed)
1632 continue;
1633
1634 if (ost->st->codec->codec_type == AVMEDIA_TYPE_AUDIO && enc->frame_size <= 1)
1635 continue;
1636 if (ost->st->codec->codec_type == AVMEDIA_TYPE_VIDEO && (os->oformat->flags & AVFMT_RAWPICTURE) && enc->codec->id == CODEC_ID_RAWVIDEO)
1637 continue;
1638
1639 for (;;) {
1640 AVPacket pkt;
1641 int fifo_bytes;
1642 av_init_packet(&pkt);
1643 pkt.data = NULL;
1644 pkt.size = 0;
1645
1646 switch (ost->st->codec->codec_type) {
1647 case AVMEDIA_TYPE_AUDIO:
1648 fifo_bytes = av_fifo_size(ost->fifo);
1649 if (fifo_bytes > 0) {
1650 /* encode any samples remaining in fifo */
1651 int frame_bytes = fifo_bytes;
1652
1653 av_fifo_generic_read(ost->fifo, audio_buf, fifo_bytes, NULL);
1654
1655 /* pad last frame with silence if needed */
1656 if (!(enc->codec->capabilities & CODEC_CAP_SMALL_LAST_FRAME)) {
1657 frame_bytes = enc->frame_size * enc->channels *
1658 av_get_bytes_per_sample(enc->sample_fmt);
1659 if (allocated_audio_buf_size < frame_bytes)
1660 exit_program(1);
1661 generate_silence(audio_buf+fifo_bytes, enc->sample_fmt, frame_bytes - fifo_bytes);
1662 }
1663 encode_audio_frame(os, ost, audio_buf, frame_bytes);
1664 } else {
1665 /* flush encoder with NULL frames until it is done
1666 returning packets */
1667 if (encode_audio_frame(os, ost, NULL, 0) == 0) {
1668 stop_encoding = 1;
1669 break;
1670 }
1671 }
1672 break;
1673 case AVMEDIA_TYPE_VIDEO:
1674 ret = avcodec_encode_video(enc, bit_buffer, bit_buffer_size, NULL);
1675 if (ret < 0) {
1676 av_log(NULL, AV_LOG_FATAL, "Video encoding failed\n");
1677 exit_program(1);
1678 }
1679 video_size += ret;
1680 if (enc->coded_frame && enc->coded_frame->key_frame)
1681 pkt.flags |= AV_PKT_FLAG_KEY;
1682 if (ost->logfile && enc->stats_out) {
1683 fprintf(ost->logfile, "%s", enc->stats_out);
1684 }
1685 if (ret <= 0) {
1686 stop_encoding = 1;
1687 break;
1688 }
1689 pkt.stream_index = ost->index;
1690 pkt.data = bit_buffer;
1691 pkt.size = ret;
1692 if (enc->coded_frame && enc->coded_frame->pts != AV_NOPTS_VALUE)
1693 pkt.pts = av_rescale_q(enc->coded_frame->pts, enc->time_base, ost->st->time_base);
1694 write_frame(os, &pkt, ost);
1695 break;
1696 default:
1697 stop_encoding = 1;
1698 }
1699 if (stop_encoding)
1700 break;
1701 }
1702 }
1703 }
1704
1705 /*
1706 * Check whether a packet from ist should be written into ost at this time
1707 */
1708 static int check_output_constraints(InputStream *ist, OutputStream *ost)
1709 {
1710 OutputFile *of = &output_files[ost->file_index];
1711 int ist_index = ist - input_streams;
1712
1713 if (ost->source_index != ist_index)
1714 return 0;
1715
1716 if (of->start_time && ist->pts < of->start_time)
1717 return 0;
1718
1719 if (of->recording_time != INT64_MAX &&
1720 av_compare_ts(ist->pts, AV_TIME_BASE_Q, of->recording_time + of->start_time,
1721 (AVRational){ 1, 1000000 }) >= 0) {
1722 ost->is_past_recording_time = 1;
1723 return 0;
1724 }
1725
1726 return 1;
1727 }
1728
1729 static void do_streamcopy(InputStream *ist, OutputStream *ost, const AVPacket *pkt)
1730 {
1731 OutputFile *of = &output_files[ost->file_index];
1732 int64_t ost_tb_start_time = av_rescale_q(of->start_time, AV_TIME_BASE_Q, ost->st->time_base);
1733 AVPacket opkt;
1734
1735 av_init_packet(&opkt);
1736
1737 if ((!ost->frame_number && !(pkt->flags & AV_PKT_FLAG_KEY)) &&
1738 !ost->copy_initial_nonkeyframes)
1739 return;
1740
1741 /* force the input stream PTS */
1742 if (ost->st->codec->codec_type == AVMEDIA_TYPE_AUDIO)
1743 audio_size += pkt->size;
1744 else if (ost->st->codec->codec_type == AVMEDIA_TYPE_VIDEO) {
1745 video_size += pkt->size;
1746 ost->sync_opts++;
1747 }
1748
1749 opkt.stream_index = ost->index;
1750 if (pkt->pts != AV_NOPTS_VALUE)
1751 opkt.pts = av_rescale_q(pkt->pts, ist->st->time_base, ost->st->time_base) - ost_tb_start_time;
1752 else
1753 opkt.pts = AV_NOPTS_VALUE;
1754
1755 if (pkt->dts == AV_NOPTS_VALUE)
1756 opkt.dts = av_rescale_q(ist->pts, AV_TIME_BASE_Q, ost->st->time_base);
1757 else
1758 opkt.dts = av_rescale_q(pkt->dts, ist->st->time_base, ost->st->time_base);
1759 opkt.dts -= ost_tb_start_time;
1760
1761 opkt.duration = av_rescale_q(pkt->duration, ist->st->time_base, ost->st->time_base);
1762 opkt.flags = pkt->flags;
1763
1764 // FIXME remove the following 2 lines they shall be replaced by the bitstream filters
1765 if ( ost->st->codec->codec_id != CODEC_ID_H264
1766 && ost->st->codec->codec_id != CODEC_ID_MPEG1VIDEO
1767 && ost->st->codec->codec_id != CODEC_ID_MPEG2VIDEO
1768 ) {
1769 if (av_parser_change(ist->st->parser, ost->st->codec, &opkt.data, &opkt.size, pkt->data, pkt->size, pkt->flags & AV_PKT_FLAG_KEY))
1770 opkt.destruct = av_destruct_packet;
1771 } else {
1772 opkt.data = pkt->data;
1773 opkt.size = pkt->size;
1774 }
1775
1776 write_frame(of->ctx, &opkt, ost);
1777 ost->st->codec->frame_number++;
1778 av_free_packet(&opkt);
1779 }
1780
1781 static void rate_emu_sleep(InputStream *ist)
1782 {
1783 if (input_files[ist->file_index].rate_emu) {
1784 int64_t pts = av_rescale(ist->pts, 1000000, AV_TIME_BASE);
1785 int64_t now = av_gettime() - ist->start;
1786 if (pts > now)
1787 usleep(pts - now);
1788 }
1789 }
1790
1791 static int transcode_audio(InputStream *ist, AVPacket *pkt, int *got_output)
1792 {
1793 AVFrame *decoded_frame;
1794 AVCodecContext *avctx = ist->st->codec;
1795 int bps = av_get_bytes_per_sample(ist->st->codec->sample_fmt);
1796 int i, ret;
1797
1798 if (!ist->decoded_frame && !(ist->decoded_frame = avcodec_alloc_frame()))
1799 return AVERROR(ENOMEM);
1800 else
1801 avcodec_get_frame_defaults(ist->decoded_frame);
1802 decoded_frame = ist->decoded_frame;
1803
1804 ret = avcodec_decode_audio4(avctx, decoded_frame, got_output, pkt);
1805 if (ret < 0) {
1806 return ret;
1807 }
1808
1809 if (!*got_output) {
1810 /* no audio frame */
1811 return ret;
1812 }
1813
1814 /* if the decoder provides a pts, use it instead of the last packet pts.
1815 the decoder could be delaying output by a packet or more. */
1816 if (decoded_frame->pts != AV_NOPTS_VALUE)
1817 ist->next_pts = decoded_frame->pts;
1818
1819 /* increment next_pts to use for the case where the input stream does not
1820 have timestamps or there are multiple frames in the packet */
1821 ist->next_pts += ((int64_t)AV_TIME_BASE * decoded_frame->nb_samples) /
1822 avctx->sample_rate;
1823
1824 // preprocess audio (volume)
1825 if (audio_volume != 256) {
1826 int decoded_data_size = decoded_frame->nb_samples * avctx->channels * bps;
1827 void *samples = decoded_frame->data[0];
1828 switch (avctx->sample_fmt) {
1829 case AV_SAMPLE_FMT_U8:
1830 {
1831 uint8_t *volp = samples;
1832 for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1833 int v = (((*volp - 128) * audio_volume + 128) >> 8) + 128;
1834 *volp++ = av_clip_uint8(v);
1835 }
1836 break;
1837 }
1838 case AV_SAMPLE_FMT_S16:
1839 {
1840 int16_t *volp = samples;
1841 for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1842 int v = ((*volp) * audio_volume + 128) >> 8;
1843 *volp++ = av_clip_int16(v);
1844 }
1845 break;
1846 }
1847 case AV_SAMPLE_FMT_S32:
1848 {
1849 int32_t *volp = samples;
1850 for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1851 int64_t v = (((int64_t)*volp * audio_volume + 128) >> 8);
1852 *volp++ = av_clipl_int32(v);
1853 }
1854 break;
1855 }
1856 case AV_SAMPLE_FMT_FLT:
1857 {
1858 float *volp = samples;
1859 float scale = audio_volume / 256.f;
1860 for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1861 *volp++ *= scale;
1862 }
1863 break;
1864 }
1865 case AV_SAMPLE_FMT_DBL:
1866 {
1867 double *volp = samples;
1868 double scale = audio_volume / 256.;
1869 for (i = 0; i < (decoded_data_size / sizeof(*volp)); i++) {
1870 *volp++ *= scale;
1871 }
1872 break;
1873 }
1874 default:
1875 av_log(NULL, AV_LOG_FATAL,
1876 "Audio volume adjustment on sample format %s is not supported.\n",
1877 av_get_sample_fmt_name(ist->st->codec->sample_fmt));
1878 exit_program(1);
1879 }
1880 }
1881
1882 rate_emu_sleep(ist);
1883
1884 for (i = 0; i < nb_output_streams; i++) {
1885 OutputStream *ost = &output_streams[i];
1886
1887 if (!check_output_constraints(ist, ost) || !ost->encoding_needed)
1888 continue;
1889 do_audio_out(output_files[ost->file_index].ctx, ost, ist, decoded_frame);
1890 }
1891
1892 return ret;
1893 }
1894
1895 static int transcode_video(InputStream *ist, AVPacket *pkt, int *got_output, int64_t *pkt_pts)
1896 {
1897 AVFrame *decoded_frame, *filtered_frame = NULL;
1898 void *buffer_to_free = NULL;
1899 int i, ret = 0;
1900 float quality;
1901 #if CONFIG_AVFILTER
1902 int frame_available = 1;
1903 #endif
1904
1905 if (!ist->decoded_frame && !(ist->decoded_frame = avcodec_alloc_frame()))
1906 return AVERROR(ENOMEM);
1907 else
1908 avcodec_get_frame_defaults(ist->decoded_frame);
1909 decoded_frame = ist->decoded_frame;
1910 pkt->pts = *pkt_pts;
1911 pkt->dts = ist->pts;
1912 *pkt_pts = AV_NOPTS_VALUE;
1913
1914 ret = avcodec_decode_video2(ist->st->codec,
1915 decoded_frame, got_output, pkt);
1916 if (ret < 0)
1917 return ret;
1918
1919 quality = same_quant ? decoded_frame->quality : 0;
1920 if (!*got_output) {
1921 /* no picture yet */
1922 return ret;
1923 }
1924 ist->next_pts = ist->pts = guess_correct_pts(&ist->pts_ctx, decoded_frame->pkt_pts,
1925 decoded_frame->pkt_dts);
1926 if (pkt->duration)
1927 ist->next_pts += av_rescale_q(pkt->duration, ist->st->time_base, AV_TIME_BASE_Q);
1928 else if (ist->st->codec->time_base.num != 0) {
1929 int ticks = ist->st->parser ? ist->st->parser->repeat_pict + 1 :
1930 ist->st->codec->ticks_per_frame;
1931 ist->next_pts += ((int64_t)AV_TIME_BASE *
1932 ist->st->codec->time_base.num * ticks) /
1933 ist->st->codec->time_base.den;
1934 }
1935 pkt->size = 0;
1936 pre_process_video_frame(ist, (AVPicture *)decoded_frame, &buffer_to_free);
1937
1938 rate_emu_sleep(ist);
1939
1940 for (i = 0; i < nb_output_streams; i++) {
1941 OutputStream *ost = &output_streams[i];
1942 int frame_size, resample_changed;
1943
1944 if (!check_output_constraints(ist, ost) || !ost->encoding_needed)
1945 continue;
1946
1947 #if CONFIG_AVFILTER
1948 resample_changed = ost->resample_width != decoded_frame->width ||
1949 ost->resample_height != decoded_frame->height ||
1950 ost->resample_pix_fmt != decoded_frame->format;
1951 if (resample_changed) {
1952 av_log(NULL, AV_LOG_INFO,
1953 "Input stream #%d:%d frame changed from size:%dx%d fmt:%s to size:%dx%d fmt:%s\n",
1954 ist->file_index, ist->st->index,
1955 ost->resample_width, ost->resample_height, av_get_pix_fmt_name(ost->resample_pix_fmt),
1956 decoded_frame->width, decoded_frame->height, av_get_pix_fmt_name(decoded_frame->format));
1957
1958 avfilter_graph_free(&ost->graph);
1959 if (configure_video_filters(ist, ost)) {
1960 av_log(NULL, AV_LOG_FATAL, "Error reinitializing filters!\n");
1961 exit_program(1);
1962 }
1963
1964 ost->resample_width = decoded_frame->width;
1965 ost->resample_height = decoded_frame->height;
1966 ost->resample_pix_fmt = decoded_frame->format;
1967 }
1968
1969 if (ist->st->sample_aspect_ratio.num)
1970 decoded_frame->sample_aspect_ratio = ist->st->sample_aspect_ratio;
1971 if (ist->st->codec->codec->capabilities & CODEC_CAP_DR1) {
1972 FrameBuffer *buf = decoded_frame->opaque;
1973 AVFilterBufferRef *fb = avfilter_get_video_buffer_ref_from_arrays(
1974 decoded_frame->data, decoded_frame->linesize,
1975 AV_PERM_READ | AV_PERM_PRESERVE,
1976 ist->st->codec->width, ist->st->codec->height,
1977 ist->st->codec->pix_fmt);
1978
1979 avfilter_copy_frame_props(fb, decoded_frame);
1980 fb->pts = ist->pts;
1981 fb->buf->priv = buf;
1982 fb->buf->free = filter_release_buffer;
1983
1984 buf->refcount++;
1985 av_buffersrc_buffer(ost->input_video_filter, fb);
1986 } else
1987 av_vsrc_buffer_add_frame(ost->input_video_filter, decoded_frame,
1988 ist->pts, decoded_frame->sample_aspect_ratio);
1989
1990 if (!ist->filtered_frame && !(ist->filtered_frame = avcodec_alloc_frame())) {
1991 av_free(buffer_to_free);
1992 return AVERROR(ENOMEM);
1993 } else
1994 avcodec_get_frame_defaults(ist->filtered_frame);
1995 filtered_frame = ist->filtered_frame;
1996
1997 frame_available = avfilter_poll_frame(ost->output_video_filter->inputs[0]);
1998 while (frame_available) {
1999 AVRational ist_pts_tb;
2000 if (ost->output_video_filter)
2001 get_filtered_video_frame(ost->output_video_filter, filtered_frame, &ost->picref, &ist_pts_tb);
2002 if (ost->picref)
2003 ist->pts = av_rescale_q(ost->picref->pts, ist_pts_tb, AV_TIME_BASE_Q);
2004 if (ost->picref->video && !ost->frame_aspect_ratio)
2005 ost->st->codec->sample_aspect_ratio = ost->picref->video->pixel_aspect;
2006 #else
2007 filtered_frame = decoded_frame;
2008 #endif
2009
2010 do_video_out(output_files[ost->file_index].ctx, ost, ist, filtered_frame, &frame_size,
2011 same_quant ? quality : ost->st->codec->global_quality);
2012 if (vstats_filename && frame_size)
2013 do_video_stats(output_files[ost->file_index].ctx, ost, frame_size);
2014 #if CONFIG_AVFILTER
2015 frame_available = ost->output_video_filter && avfilter_poll_frame(ost->output_video_filter->inputs[0]);
2016 if (ost->picref)
2017 avfilter_unref_buffer(ost->picref);
2018 }
2019 #endif
2020 }
2021
2022 av_free(buffer_to_free);
2023 return ret;
2024 }
2025
2026 static int transcode_subtitles(InputStream *ist, AVPacket *pkt, int *got_output)
2027 {
2028 AVSubtitle subtitle;
2029 int i, ret = avcodec_decode_subtitle2(ist->st->codec,
2030 &subtitle, got_output, pkt);
2031 if (ret < 0)
2032 return ret;
2033 if (!*got_output)
2034 return ret;
2035
2036 rate_emu_sleep(ist);
2037
2038 for (i = 0; i < nb_output_streams; i++) {
2039 OutputStream *ost = &output_streams[i];
2040
2041 if (!check_output_constraints(ist, ost) || !ost->encoding_needed)
2042 continue;
2043
2044 do_subtitle_out(output_files[ost->file_index].ctx, ost, ist, &subtitle, pkt->pts);
2045 }
2046
2047 avsubtitle_free(&subtitle);
2048 return ret;
2049 }
2050
2051 /* pkt = NULL means EOF (needed to flush decoder buffers) */
2052 static int output_packet(InputStream *ist,
2053 OutputStream *ost_table, int nb_ostreams,
2054 const AVPacket *pkt)
2055 {
2056 int i;
2057 int got_output;
2058 int64_t pkt_pts = AV_NOPTS_VALUE;
2059 AVPacket avpkt;
2060
2061 if (ist->next_pts == AV_NOPTS_VALUE)
2062 ist->next_pts = ist->pts;
2063
2064 if (pkt == NULL) {
2065 /* EOF handling */
2066 av_init_packet(&avpkt);
2067 avpkt.data = NULL;
2068 avpkt.size = 0;
2069 goto handle_eof;
2070 } else {
2071 avpkt = *pkt;
2072 }
2073
2074 if (pkt->dts != AV_NOPTS_VALUE)
2075 ist->next_pts = ist->pts = av_rescale_q(pkt->dts, ist->st->time_base, AV_TIME_BASE_Q);
2076 if (pkt->pts != AV_NOPTS_VALUE)
2077 pkt_pts = av_rescale_q(pkt->pts, ist->st->time_base, AV_TIME_BASE_Q);
2078
2079 // while we have more to decode or while the decoder did output something on EOF
2080 while (ist->decoding_needed && (avpkt.size > 0 || (!pkt && got_output))) {
2081 int ret = 0;
2082 handle_eof:
2083
2084 ist->pts = ist->next_pts;
2085
2086 if (avpkt.size && avpkt.size != pkt->size) {
2087 av_log(NULL, ist->showed_multi_packet_warning ? AV_LOG_VERBOSE : AV_LOG_WARNING,
2088 "Multiple frames in a packet from stream %d\n", pkt->stream_index);
2089 ist->showed_multi_packet_warning = 1;
2090 }
2091
2092 switch (ist->st->codec->codec_type) {
2093 case AVMEDIA_TYPE_AUDIO:
2094 ret = transcode_audio (ist, &avpkt, &got_output);
2095 break;
2096 case AVMEDIA_TYPE_VIDEO:
2097 ret = transcode_video (ist, &avpkt, &got_output, &pkt_pts);
2098 break;
2099 case AVMEDIA_TYPE_SUBTITLE:
2100 ret = transcode_subtitles(ist, &avpkt, &got_output);
2101 break;
2102 default:
2103 return -1;
2104 }
2105
2106 if (ret < 0)
2107 return ret;
2108 // touch data and size only if not EOF
2109 if (pkt) {
2110 avpkt.data += ret;
2111 avpkt.size -= ret;
2112 }
2113 if (!got_output) {
2114 continue;
2115 }
2116 }
2117
2118 /* handle stream copy */
2119 if (!ist->decoding_needed) {
2120 rate_emu_sleep(ist);
2121 ist->pts = ist->next_pts;
2122 switch (ist->st->codec->codec_type) {
2123 case AVMEDIA_TYPE_AUDIO:
2124 ist->next_pts += ((int64_t)AV_TIME_BASE * ist->st->codec->frame_size) /
2125 ist->st->codec->sample_rate;
2126 break;
2127 case AVMEDIA_TYPE_VIDEO:
2128 if (ist->st->codec->time_base.num != 0) {
2129 int ticks = ist->st->parser ? ist->st->parser->repeat_pict + 1 : ist->st->codec->ticks_per_frame;
2130 ist->next_pts += ((int64_t)AV_TIME_BASE *
2131 ist->st->codec->time_base.num * ticks) /
2132 ist->st->codec->time_base.den;
2133 }
2134 break;
2135 }
2136 }
2137 for (i = 0; pkt && i < nb_ostreams; i++) {
2138 OutputStream *ost = &ost_table[i];
2139
2140 if (!check_output_constraints(ist, ost) || ost->encoding_needed)
2141 continue;
2142
2143 do_streamcopy(ist, ost, pkt);
2144 }
2145
2146 return 0;
2147 }
2148
2149 static void print_sdp(OutputFile *output_files, int n)
2150 {
2151 char sdp[2048];
2152 int i;
2153 AVFormatContext **avc = av_malloc(sizeof(*avc) * n);
2154
2155 if (!avc)
2156 exit_program(1);
2157 for (i = 0; i < n; i++)
2158 avc[i] = output_files[i].ctx;
2159
2160 av_sdp_create(avc, n, sdp, sizeof(sdp));
2161 printf("SDP:\n%s\n", sdp);
2162 fflush(stdout);
2163 av_freep(&avc);
2164 }
2165
2166 static int init_input_stream(int ist_index, OutputStream *output_streams, int nb_output_streams,
2167 char *error, int error_len)
2168 {
2169 int i;
2170 InputStream *ist = &input_streams[ist_index];
2171 if (ist->decoding_needed) {
2172 AVCodec *codec = ist->dec;
2173 if (!codec) {
2174 snprintf(error, error_len, "Decoder (codec id %d) not found for input stream #%d:%d",
2175 ist->st->codec->codec_id, ist->file_index, ist->st->index);
2176 return AVERROR(EINVAL);
2177 }
2178
2179 /* update requested sample format for the decoder based on the
2180 corresponding encoder sample format */
2181 for (i = 0; i < nb_output_streams; i++) {
2182 OutputStream *ost = &output_streams[i];
2183 if (ost->source_index == ist_index) {
2184 update_sample_fmt(ist->st->codec, codec, ost->st->codec);
2185 break;
2186 }
2187 }
2188
2189 if (codec->type == AVMEDIA_TYPE_VIDEO && codec->capabilities & CODEC_CAP_DR1) {
2190 ist->st->codec->get_buffer = codec_get_buffer;
2191 ist->st->codec->release_buffer = codec_release_buffer;
2192 ist->st->codec->opaque = ist;
2193 }
2194
2195 if (avcodec_open2(ist->st->codec, codec, &ist->opts) < 0) {
2196 snprintf(error, error_len, "Error while opening decoder for input stream #%d:%d",
2197 ist->file_index, ist->st->index);
2198 return AVERROR(EINVAL);
2199 }
2200 assert_codec_experimental(ist->st->codec, 0);
2201 assert_avoptions(ist->opts);
2202 }
2203
2204 ist->pts = ist->st->avg_frame_rate.num ? - ist->st->codec->has_b_frames * AV_TIME_BASE / av_q2d(ist->st->avg_frame_rate) : 0;
2205 ist->next_pts = AV_NOPTS_VALUE;
2206 init_pts_correction(&ist->pts_ctx);
2207 ist->is_start = 1;
2208
2209 return 0;
2210 }
2211
2212 static int transcode_init(OutputFile *output_files,
2213 int nb_output_files,
2214 InputFile *input_files,
2215 int nb_input_files)
2216 {
2217 int ret = 0, i, j, k;
2218 AVFormatContext *oc;
2219 AVCodecContext *codec, *icodec;
2220 OutputStream *ost;
2221 InputStream *ist;
2222 char error[1024];
2223 int want_sdp = 1;
2224
2225 /* init framerate emulation */
2226 for (i = 0; i < nb_input_files; i++) {
2227 InputFile *ifile = &input_files[i];
2228 if (ifile->rate_emu)
2229 for (j = 0; j < ifile->nb_streams; j++)
2230 input_streams[j + ifile->ist_index].start = av_gettime();
2231 }
2232
2233 /* output stream init */
2234 for (i = 0; i < nb_output_files; i++) {
2235 oc = output_files[i].ctx;
2236 if (!oc->nb_streams && !(oc->oformat->flags & AVFMT_NOSTREAMS)) {
2237 av_dump_format(oc, i, oc->filename, 1);
2238 av_log(NULL, AV_LOG_ERROR, "Output file #%d does not contain any stream\n", i);
2239 return AVERROR(EINVAL);
2240 }
2241 }
2242
2243 /* for each output stream, we compute the right encoding parameters */
2244 for (i = 0; i < nb_output_streams; i++) {
2245 ost = &output_streams[i];
2246 oc = output_files[ost->file_index].ctx;
2247 ist = &input_streams[ost->source_index];
2248
2249 if (ost->attachment_filename)
2250 continue;
2251
2252 codec = ost->st->codec;
2253 icodec = ist->st->codec;
2254
2255 ost->st->disposition = ist->st->disposition;
2256 codec->bits_per_raw_sample = icodec->bits_per_raw_sample;
2257 codec->chroma_sample_location = icodec->chroma_sample_location;
2258
2259 if (ost->stream_copy) {
2260 uint64_t extra_size = (uint64_t)icodec->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE;
2261
2262 if (extra_size > INT_MAX) {
2263 return AVERROR(EINVAL);
2264 }
2265
2266 /* if stream_copy is selected, no need to decode or encode */
2267 codec->codec_id = icodec->codec_id;
2268 codec->codec_type = icodec->codec_type;
2269
2270 if (!codec->codec_tag) {
2271 if (!oc->oformat->codec_tag ||
2272 av_codec_get_id (oc->oformat->codec_tag, icodec->codec_tag) == codec->codec_id ||
2273 av_codec_get_tag(oc->oformat->codec_tag, icodec->codec_id) <= 0)
2274 codec->codec_tag = icodec->codec_tag;
2275 }
2276
2277 codec->bit_rate = icodec->bit_rate;
2278 codec->rc_max_rate = icodec->rc_max_rate;
2279 codec->rc_buffer_size = icodec->rc_buffer_size;
2280 codec->field_order = icodec->field_order;
2281 codec->extradata = av_mallocz(extra_size);
2282 if (!codec->extradata) {
2283 return AVERROR(ENOMEM);
2284 }
2285 memcpy(codec->extradata, icodec->extradata, icodec->extradata_size);
2286 codec->extradata_size = icodec->extradata_size;
2287 if (!copy_tb) {
2288 codec->time_base = icodec->time_base;
2289 codec->time_base.num *= icodec->ticks_per_frame;
2290 av_reduce(&codec->time_base.num, &codec->time_base.den,
2291 codec->time_base.num, codec->time_base.den, INT_MAX);
2292 } else
2293 codec->time_base = ist->st->time_base;
2294
2295 switch (codec->codec_type) {
2296 case AVMEDIA_TYPE_AUDIO:
2297 if (audio_volume != 256) {
2298 av_log(NULL, AV_LOG_FATAL, "-acodec copy and -vol are incompatible (frames are not decoded)\n");
2299 exit_program(1);
2300 }
2301 codec->channel_layout = icodec->channel_layout;
2302 codec->sample_rate = icodec->sample_rate;
2303 codec->channels = icodec->channels;
2304 codec->frame_size = icodec->frame_size;
2305 codec->audio_service_type = icodec->audio_service_type;
2306 codec->block_align = icodec->block_align;
2307 break;
2308 case AVMEDIA_TYPE_VIDEO:
2309 codec->pix_fmt = icodec->pix_fmt;
2310 codec->width = icodec->width;
2311 codec->height = icodec->height;
2312 codec->has_b_frames = icodec->has_b_frames;
2313 if (!codec->sample_aspect_ratio.num) {
2314 codec->sample_aspect_ratio =
2315 ost->st->sample_aspect_ratio =
2316 ist->st->sample_aspect_ratio.num ? ist->st->sample_aspect_ratio :
2317 ist->st->codec->sample_aspect_ratio.num ?
2318 ist->st->codec->sample_aspect_ratio : (AVRational){0, 1};
2319 }
2320 break;
2321 case AVMEDIA_TYPE_SUBTITLE:
2322 codec->width = icodec->width;
2323 codec->height = icodec->height;
2324 break;
2325 case AVMEDIA_TYPE_DATA:
2326 case AVMEDIA_TYPE_ATTACHMENT:
2327 break;
2328 default:
2329 abort();
2330 }
2331 } else {
2332 if (!ost->enc)
2333 ost->enc = avcodec_find_encoder(ost->st->codec->codec_id);
2334
2335 ist->decoding_needed = 1;
2336 ost->encoding_needed = 1;
2337
2338 switch (codec->codec_type) {
2339 case AVMEDIA_TYPE_AUDIO:
2340 ost->fifo = av_fifo_alloc(1024);
2341 if (!ost->fifo) {
2342 return AVERROR(ENOMEM);
2343 }
2344 ost->reformat_pair = MAKE_SFMT_PAIR(AV_SAMPLE_FMT_NONE,AV_SAMPLE_FMT_NONE);
2345
2346 if (!codec->sample_rate)
2347 codec->sample_rate = icodec->sample_rate;
2348 choose_sample_rate(ost->st, ost->enc);
2349 codec->time_base = (AVRational){ 1, codec->sample_rate };
2350
2351 if (codec->sample_fmt == AV_SAMPLE_FMT_NONE)
2352 codec->sample_fmt = icodec->sample_fmt;
2353 choose_sample_fmt(ost->st, ost->enc);
2354
2355 if (!codec->channels)
2356 codec->channels = icodec->channels;
2357 codec->channel_layout = icodec->channel_layout;
2358 if (av_get_channel_layout_nb_channels(codec->channel_layout) != codec->channels)
2359 codec->channel_layout = 0;
2360
2361 ost->audio_resample = codec-> sample_rate != icodec->sample_rate || audio_sync_method > 1;
2362 icodec->request_channels = codec-> channels;
2363 ost->resample_sample_fmt = icodec->sample_fmt;
2364 ost->resample_sample_rate = icodec->sample_rate;
2365 ost->resample_channels = icodec->channels;
2366 break;
2367 case AVMEDIA_TYPE_VIDEO:
2368 if (codec->pix_fmt == PIX_FMT_NONE)
2369 codec->pix_fmt = icodec->pix_fmt;
2370 choose_pixel_fmt(ost->st, ost->enc);
2371
2372 if (ost->st->codec->pix_fmt == PIX_FMT_NONE) {
2373 av_log(NULL, AV_LOG_FATAL, "Video pixel format is unknown, stream cannot be encoded\n");
2374 exit_program(1);
2375 }
2376
2377 if (!codec->width || !codec->height) {
2378 codec->width = icodec->width;
2379 codec->height = icodec->height;
2380 }
2381
2382 ost->video_resample = codec->width != icodec->width ||
2383 codec->height != icodec->height ||
2384 codec->pix_fmt != icodec->pix_fmt;
2385 if (ost->video_resample) {
2386 #if !CONFIG_AVFILTER
2387 avcodec_get_frame_defaults(&ost->pict_tmp);
2388 if (avpicture_alloc((AVPicture*)&ost->pict_tmp, codec->pix_fmt,
2389 codec->width, codec->height)) {
2390 av_log(NULL, AV_LOG_FATAL, "Cannot allocate temp picture, check pix fmt\n");
2391 exit_program(1);
2392 }
2393 ost->img_resample_ctx = sws_getContext(
2394 icodec->width,
2395 icodec->height,
2396 icodec->pix_fmt,
2397 codec->width,
2398 codec->height,
2399 codec->pix_fmt,
2400 ost->sws_flags, NULL, NULL, NULL);
2401 if (ost->img_resample_ctx == NULL) {
2402 av_log(NULL, AV_LOG_FATAL, "Cannot get resampling context\n");
2403 exit_program(1);
2404 }
2405 #endif
2406 codec->bits_per_raw_sample = 0;
2407 }
2408
2409 ost->resample_height = icodec->height;
2410 ost->resample_width = icodec->width;
2411 ost->resample_pix_fmt = icodec->pix_fmt;
2412
2413 if (!ost->frame_rate.num)
2414 ost->frame_rate = ist->st->r_frame_rate.num ? ist->st->r_frame_rate : (AVRational) { 25, 1 };
2415 if (ost->enc && ost->enc->supported_framerates && !ost->force_fps) {
2416 int idx = av_find_nearest_q_idx(ost->frame_rate, ost->enc->supported_framerates);
2417 ost->frame_rate = ost->enc->supported_framerates[idx];
2418 }
2419 codec->time_base = (AVRational){ost->frame_rate.den, ost->frame_rate.num};
2420
2421 #if CONFIG_AVFILTER
2422 if (configure_video_filters(ist, ost)) {
2423 av_log(NULL, AV_LOG_FATAL, "Error opening filters!\n");
2424 exit(1);
2425 }
2426 #endif
2427 break;
2428 case AVMEDIA_TYPE_SUBTITLE:
2429 break;
2430 default:
2431 abort();
2432 break;
2433 }
2434 /* two pass mode */
2435 if ((codec->flags & (CODEC_FLAG_PASS1 | CODEC_FLAG_PASS2))) {
2436 char logfilename[1024];
2437 FILE *f;
2438
2439 snprintf(logfilename, sizeof(logfilename), "%s-%d.log",
2440 pass_logfilename_prefix ? pass_logfilename_prefix : DEFAULT_PASS_LOGFILENAME_PREFIX,
2441 i);
2442 if (codec->flags & CODEC_FLAG_PASS1) {
2443 f = fopen(logfilename, "wb");
2444 if (!f) {
2445 av_log(NULL, AV_LOG_FATAL, "Cannot write log file '%s' for pass-1 encoding: %s\n",
2446 logfilename, strerror(errno));
2447 exit_program(1);
2448 }
2449 ost->logfile = f;
2450 } else {
2451 char *logbuffer;
2452 size_t logbuffer_size;
2453 if (cmdutils_read_file(logfilename, &logbuffer, &logbuffer_size) < 0) {
2454 av_log(NULL, AV_LOG_FATAL, "Error reading log file '%s' for pass-2 encoding\n",
2455 logfilename);
2456 exit_program(1);
2457 }
2458 codec->stats_in = logbuffer;
2459 }
2460 }
2461 }
2462 if (codec->codec_type == AVMEDIA_TYPE_VIDEO) {
2463 int size = codec->width * codec->height;
2464 bit_buffer_size = FFMAX(bit_buffer_size, 6 * size + 200);
2465 }
2466 }
2467
2468 if (!bit_buffer)
2469 bit_buffer = av_malloc(bit_buffer_size);
2470 if (!bit_buffer) {
2471 av_log(NULL, AV_LOG_ERROR, "Cannot allocate %d bytes output buffer\n",
2472 bit_buffer_size);
2473 return AVERROR(ENOMEM);
2474 }
2475
2476 /* open each encoder */
2477 for (i = 0; i < nb_output_streams; i++) {
2478 ost = &output_streams[i];
2479 if (ost->encoding_needed) {
2480 AVCodec *codec = ost->enc;
2481 AVCodecContext *dec = input_streams[ost->source_index].st->codec;
2482 if (!codec) {
2483 snprintf(error, sizeof(error), "Encoder (codec id %d) not found for output stream #%d:%d",
2484 ost->st->codec->codec_id, ost->file_index, ost->index);
2485 ret = AVERROR(EINVAL);
2486 goto dump_format;
2487 }
2488 if (dec->subtitle_header) {
2489 ost->st->codec->subtitle_header = av_malloc(dec->subtitle_header_size);
2490 if (!ost->st->codec->subtitle_header) {
2491 ret = AVERROR(ENOMEM);
2492 goto dump_format;
2493 }
2494 memcpy(ost->st->codec->subtitle_header, dec->subtitle_header, dec->subtitle_header_size);
2495 ost->st->codec->subtitle_header_size = dec->subtitle_header_size;
2496 }
2497 if (avcodec_open2(ost->st->codec, codec, &ost->opts) < 0) {
2498 snprintf(error, sizeof(error), "Error while opening encoder for output stream #%d:%d - maybe incorrect parameters such as bit_rate, rate, width or height",
2499 ost->file_index, ost->index);
2500 ret = AVERROR(EINVAL);
2501 goto dump_format;
2502 }
2503 assert_codec_experimental(ost->st->codec, 1);
2504 assert_avoptions(ost->opts);
2505 if (ost->st->codec->bit_rate && ost->st->codec->bit_rate < 1000)
2506 av_log(NULL, AV_LOG_WARNING, "The bitrate parameter is set too low."
2507 "It takes bits/s as argument, not kbits/s\n");
2508 extra_size += ost->st->codec->extradata_size;
2509
2510 if (ost->st->codec->me_threshold)
2511 input_streams[ost->source_index].st->codec->debug |= FF_DEBUG_MV;
2512 }
2513 }
2514
2515 /* init input streams */
2516 for (i = 0; i < nb_input_streams; i++)
2517 if ((ret = init_input_stream(i, output_streams, nb_output_streams, error, sizeof(error))) < 0)
2518 goto dump_format;
2519
2520 /* discard unused programs */
2521 for (i = 0; i < nb_input_files; i++) {
2522 InputFile *ifile = &input_files[i];
2523 for (j = 0; j < ifile->ctx->nb_programs; j++) {
2524 AVProgram *p = ifile->ctx->programs[j];
2525 int discard = AVDISCARD_ALL;
2526
2527 for (k = 0; k < p->nb_stream_indexes; k++)
2528 if (!input_streams[ifile->ist_index + p->stream_index[k]].discard) {
2529 discard = AVDISCARD_DEFAULT;
2530 break;
2531 }
2532 p->discard = discard;
2533 }
2534 }
2535
2536 /* open files and write file headers */
2537 for (i = 0; i < nb_output_files; i++) {
2538 oc = output_files[i].ctx;
2539 oc->interrupt_callback = int_cb;
2540 if (avformat_write_header(oc, &output_files[i].opts) < 0) {
2541 snprintf(error, sizeof(error), "Could not write header for output file #%d (incorrect codec parameters ?)", i);
2542 ret = AVERROR(EINVAL);
2543 goto dump_format;
2544 }
2545 assert_avoptions(output_files[i].opts);
2546 if (strcmp(oc->oformat->name, "rtp")) {
2547 want_sdp = 0;
2548 }
2549 }
2550
2551 dump_format:
2552 /* dump the file output parameters - cannot be done before in case
2553 of stream copy */
2554 for (i = 0; i < nb_output_files; i++) {
2555 av_dump_format(output_files[i].ctx, i, output_files[i].ctx->filename, 1);
2556 }
2557
2558 /* dump the stream mapping */
2559 av_log(NULL, AV_LOG_INFO, "Stream mapping:\n");
2560 for (i = 0; i < nb_output_streams; i++) {
2561 ost = &output_streams[i];
2562
2563 if (ost->attachment_filename) {
2564 /* an attached file */
2565 av_log(NULL, AV_LOG_INFO, " File %s -> Stream #%d:%d\n",
2566 ost->attachment_filename, ost->file_index, ost->index);
2567 continue;
2568 }
2569 av_log(NULL, AV_LOG_INFO, " Stream #%d:%d -> #%d:%d",
2570 input_streams[ost->source_index].file_index,
2571 input_streams[ost->source_index].st->index,
2572 ost->file_index,
2573 ost->index);
2574 if (ost->sync_ist != &input_streams[ost->source_index])
2575 av_log(NULL, AV_LOG_INFO, " [sync #%d:%d]",
2576 ost->sync_ist->file_index,
2577 ost->sync_ist->st->index);
2578 if (ost->stream_copy)
2579 av_log(NULL, AV_LOG_INFO, " (copy)");
2580 else
2581 av_log(NULL, AV_LOG_INFO, " (%s -> %s)", input_streams[ost->source_index].dec ?
2582 input_streams[ost->source_index].dec->name : "?",
2583 ost->enc ? ost->enc->name : "?");
2584 av_log(NULL, AV_LOG_INFO, "\n");
2585 }
2586
2587 if (ret) {
2588 av_log(NULL, AV_LOG_ERROR, "%s\n", error);
2589 return ret;
2590 }
2591
2592 if (want_sdp) {
2593 print_sdp(output_files, nb_output_files);
2594 }
2595
2596 return 0;
2597 }
2598
2599 /*
2600 * The following code is the main loop of the file converter
2601 */
2602 static int transcode(OutputFile *output_files,
2603 int nb_output_files,
2604 InputFile *input_files,
2605 int nb_input_files)
2606 {
2607 int ret, i;
2608 AVFormatContext *is, *os;
2609 OutputStream *ost;
2610 InputStream *ist;
2611 uint8_t *no_packet;
2612 int no_packet_count = 0;
2613 int64_t timer_start;
2614
2615 if (!(no_packet = av_mallocz(nb_input_files)))
2616 exit_program(1);
2617
2618 ret = transcode_init(output_files, nb_output_files, input_files, nb_input_files);
2619 if (ret < 0)
2620 goto fail;
2621
2622 av_log(NULL, AV_LOG_INFO, "Press ctrl-c to stop encoding\n");
2623 term_init();
2624
2625 timer_start = av_gettime();
2626
2627 for (; received_sigterm == 0;) {
2628 int file_index, ist_index;
2629 AVPacket pkt;
2630 int64_t ipts_min;
2631 double opts_min;
2632
2633 ipts_min = INT64_MAX;
2634 opts_min = 1e100;
2635
2636 /* select the stream that we must read now by looking at the
2637 smallest output pts */
2638 file_index = -1;
2639 for (i = 0; i < nb_output_streams; i++) {
2640 OutputFile *of;
2641 int64_t ipts;
2642 double opts;
2643 ost = &output_streams[i];
2644 of = &output_files[ost->file_index];
2645 os = output_files[ost->file_index].ctx;
2646 ist = &input_streams[ost->source_index];
2647 if (ost->is_past_recording_time || no_packet[ist->file_index] ||
2648 (os->pb && avio_tell(os->pb) >= of->limit_filesize))
2649 continue;
2650 opts = ost->st->pts.val * av_q2d(ost->st->time_base);
2651 ipts = ist->pts;
2652 if (!input_files[ist->file_index].eof_reached) {
2653 if (ipts < ipts_min) {
2654 ipts_min = ipts;
2655 if (input_sync)
2656 file_index = ist->file_index;
2657 }
2658 if (opts < opts_min) {
2659 opts_min = opts;
2660 if (!input_sync) file_index = ist->file_index;
2661 }
2662 }
2663 if (ost->frame_number >= ost->max_frames) {
2664 int j;
2665 for (j = 0; j < of->ctx->nb_streams; j++)
2666 output_streams[of->ost_index + j].is_past_recording_time = 1;
2667 continue;
2668 }
2669 }
2670 /* if none, if is finished */
2671 if (file_index < 0) {
2672 if (no_packet_count) {
2673 no_packet_count = 0;
2674 memset(no_packet, 0, nb_input_files);
2675 usleep(10000);
2676 continue;
2677 }
2678 break;
2679 }
2680
2681 /* read a frame from it and output it in the fifo */
2682 is = input_files[file_index].ctx;
2683 ret = av_read_frame(is, &pkt);
2684 if (ret == AVERROR(EAGAIN)) {
2685 no_packet[file_index] = 1;
2686 no_packet_count++;
2687 continue;
2688 }
2689 if (ret < 0) {
2690 input_files[file_index].eof_reached = 1;
2691 if (opt_shortest)
2692 break;
2693 else
2694 continue;
2695 }
2696
2697 no_packet_count = 0;
2698 memset(no_packet, 0, nb_input_files);
2699
2700 if (do_pkt_dump) {
2701 av_pkt_dump_log2(NULL, AV_LOG_DEBUG, &pkt, do_hex_dump,
2702 is->streams[pkt.stream_index]);
2703 }
2704 /* the following test is needed in case new streams appear
2705 dynamically in stream : we ignore them */
2706 if (pkt.stream_index >= input_files[file_index].nb_streams)
2707 goto discard_packet;
2708 ist_index = input_files[file_index].ist_index + pkt.stream_index;
2709 ist = &input_streams[ist_index];
2710 if (ist->discard)
2711 goto discard_packet;
2712
2713 if (pkt.dts != AV_NOPTS_VALUE)
2714 pkt.dts += av_rescale_q(input_files[ist->file_index].ts_offset, AV_TIME_BASE_Q, ist->st->time_base);
2715 if (pkt.pts != AV_NOPTS_VALUE)
2716 pkt.pts += av_rescale_q(input_files[ist->file_index].ts_offset, AV_TIME_BASE_Q, ist->st->time_base);
2717
2718 if (pkt.pts != AV_NOPTS_VALUE)
2719 pkt.pts *= ist->ts_scale;
2720 if (pkt.dts != AV_NOPTS_VALUE)
2721 pkt.dts *= ist->ts_scale;
2722
2723 //fprintf(stderr, "next:%"PRId64" dts:%"PRId64" off:%"PRId64" %d\n",
2724 // ist->next_pts,
2725 // pkt.dts, input_files[ist->file_index].ts_offset,
2726 // ist->st->codec->codec_type);
2727 if (pkt.dts != AV_NOPTS_VALUE && ist->next_pts != AV_NOPTS_VALUE
2728 && (is->iformat->flags & AVFMT_TS_DISCONT)) {
2729 int64_t pkt_dts = av_rescale_q(pkt.dts, ist->st->time_base, AV_TIME_BASE_Q);
2730 int64_t delta = pkt_dts - ist->next_pts;
2731 if ((FFABS(delta) > 1LL * dts_delta_threshold * AV_TIME_BASE || pkt_dts + 1 < ist->pts) && !copy_ts) {
2732 input_files[ist->file_index].ts_offset -= delta;
2733 av_log(NULL, AV_LOG_DEBUG,
2734 "timestamp discontinuity %"PRId64", new offset= %"PRId64"\n",
2735 delta, input_files[ist->file_index].ts_offset);
2736 pkt.dts-= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);
2737 if (pkt.pts != AV_NOPTS_VALUE)
2738 pkt.pts-= av_rescale_q(delta, AV_TIME_BASE_Q, ist->st->time_base);
2739 }
2740 }
2741
2742 // fprintf(stderr,"read #%d.%d size=%d\n", ist->file_index, ist->st->index, pkt.size);
2743 if (output_packet(ist, output_streams, nb_output_streams, &pkt) < 0) {
2744
2745 av_log(NULL, AV_LOG_ERROR, "Error while decoding stream #%d:%d\n",
2746 ist->file_index, ist->st->index);
2747 if (exit_on_error)
2748 exit_program(1);
2749 av_free_packet(&pkt);
2750 continue;
2751 }
2752
2753 discard_packet:
2754 av_free_packet(&pkt);
2755
2756 /* dump report by using the output first video and audio streams */
2757 print_report(output_files, output_streams, nb_output_streams, 0, timer_start);
2758 }
2759
2760 /* at the end of stream, we must flush the decoder buffers */
2761 for (i = 0; i < nb_input_streams; i++) {
2762 ist = &input_streams[i];
2763 if (ist->decoding_needed) {
2764 output_packet(ist, output_streams, nb_output_streams, NULL);
2765 }
2766 }
2767 flush_encoders(output_streams, nb_output_streams);
2768
2769 term_exit();
2770
2771 /* write the trailer if needed and close file */
2772 for (i = 0; i < nb_output_files; i++) {
2773 os = output_files[i].ctx;
2774 av_write_trailer(os);
2775 }
2776
2777 /* dump report by using the first video and audio streams */
2778 print_report(output_files, output_streams, nb_output_streams, 1, timer_start);
2779
2780 /* close each encoder */
2781 for (i = 0; i < nb_output_streams; i++) {
2782 ost = &output_streams[i];
2783 if (ost->encoding_needed) {
2784 av_freep(&ost->st->codec->stats_in);
2785 avcodec_close(ost->st->codec);
2786 }
2787 #if CONFIG_AVFILTER
2788 avfilter_graph_free(&ost->graph);
2789 #endif
2790 }
2791
2792 /* close each decoder */
2793 for (i = 0; i < nb_input_streams; i++) {
2794 ist = &input_streams[i];
2795 if (ist->decoding_needed) {
2796 avcodec_close(ist->st->codec);
2797 }
2798 }
2799
2800 /* finished ! */
2801 ret = 0;
2802
2803 fail:
2804 av_freep(&bit_buffer);
2805 av_freep(&no_packet);
2806
2807 if (output_streams) {
2808 for (i = 0; i < nb_output_streams; i++) {
2809 ost = &output_streams[i];
2810 if (ost) {
2811 if (ost->stream_copy)
2812 av_freep(&ost->st->codec->extradata);
2813 if (ost->logfile) {
2814 fclose(ost->logfile);
2815 ost->logfile = NULL;
2816 }
2817 av_fifo_free(ost->fifo); /* works even if fifo is not
2818 initialized but set to zero */
2819 av_freep(&ost->st->codec->subtitle_header);
2820 av_free(ost->pict_tmp.data[0]);
2821 av_free(ost->forced_kf_pts);
2822 if (ost->video_resample)
2823 sws_freeContext(ost->img_resample_ctx);
2824 if (ost->resample)
2825 audio_resample_close(ost->resample);
2826 if (ost->reformat_ctx)
2827 av_audio_convert_free(ost->reformat_ctx);
2828 av_dict_free(&ost->opts);
2829 }
2830 }
2831 }
2832 return ret;
2833 }
2834
2835 static double parse_frame_aspect_ratio(const char *arg)
2836 {
2837 int x = 0, y = 0;
2838 double ar = 0;
2839 const char *p;
2840 char *end;
2841
2842 p = strchr(arg, ':');
2843 if (p) {
2844 x = strtol(arg, &end, 10);
2845 if (end == p)
2846 y = strtol(end + 1, &end, 10);
2847 if (x > 0 && y > 0)
2848 ar = (double)x / (double)y;
2849 } else
2850 ar = strtod(arg, NULL);
2851
2852 if (!ar) {
2853 av_log(NULL, AV_LOG_FATAL, "Incorrect aspect ratio specification.\n");
2854 exit_program(1);
2855 }
2856 return ar;
2857 }
2858
2859 static int opt_audio_codec(OptionsContext *o, const char *opt, const char *arg)
2860 {
2861 return parse_option(o, "codec:a", arg, options);
2862 }
2863
2864 static int opt_video_codec(OptionsContext *o, const char *opt, const char *arg)
2865 {
2866 return parse_option(o, "codec:v", arg, options);
2867 }
2868
2869 static int opt_subtitle_codec(OptionsContext *o, const char *opt, const char *arg)
2870 {
2871 return parse_option(o, "codec:s", arg, options);
2872 }
2873
2874 static int opt_data_codec(OptionsContext *o, const char *opt, const char *arg)
2875 {
2876 return parse_option(o, "codec:d", arg, options);
2877 }
2878
2879 static int opt_map(OptionsContext *o, const char *opt, const char *arg)
2880 {
2881 StreamMap *m = NULL;
2882 int i, negative = 0, file_idx;
2883 int sync_file_idx = -1, sync_stream_idx;
2884 char *p, *sync;
2885 char *map;
2886
2887 if (*arg == '-') {
2888 negative = 1;
2889 arg++;
2890 }
2891 map = av_strdup(arg);
2892
2893 /* parse sync stream first, just pick first matching stream */
2894 if (sync = strchr(map, ',')) {
2895 *sync = 0;
2896 sync_file_idx = strtol(sync + 1, &sync, 0);
2897 if (sync_file_idx >= nb_input_files || sync_file_idx < 0) {
2898 av_log(NULL, AV_LOG_FATAL, "Invalid sync file index: %d.\n", sync_file_idx);
2899 exit_program(1);
2900 }
2901 if (*sync)
2902 sync++;
2903 for (i = 0; i < input_files[sync_file_idx].nb_streams; i++)
2904 if (check_stream_specifier(input_files[sync_file_idx].ctx,
2905 input_files[sync_file_idx].ctx->streams[i], sync) == 1) {
2906 sync_stream_idx = i;
2907 break;
2908 }
2909 if (i == input_files[sync_file_idx].nb_streams) {
2910 av_log(NULL, AV_LOG_FATAL, "Sync stream specification in map %s does not "
2911 "match any streams.\n", arg);
2912 exit_program(1);
2913 }
2914 }
2915
2916
2917 file_idx = strtol(map, &p, 0);
2918 if (file_idx >= nb_input_files || file_idx < 0) {
2919 av_log(NULL, AV_LOG_FATAL, "Invalid input file index: %d.\n", file_idx);
2920 exit_program(1);
2921 }
2922 if (negative)
2923 /* disable some already defined maps */
2924 for (i = 0; i < o->nb_stream_maps; i++) {
2925 m = &o->stream_maps[i];
2926 if (file_idx == m->file_index &&
2927 check_stream_specifier(input_files[m->file_index].ctx,
2928 input_files[m->file_index].ctx->streams[m->stream_index],
2929 *p == ':' ? p + 1 : p) > 0)
2930 m->disabled = 1;
2931 }
2932 else
2933 for (i = 0; i < input_files[file_idx].nb_streams; i++) {
2934 if (check_stream_specifier(input_files[file_idx].ctx, input_files[file_idx].ctx->streams[i],
2935 *p == ':' ? p + 1 : p) <= 0)
2936 continue;
2937 o->stream_maps = grow_array(o->stream_maps, sizeof(*o->stream_maps),
2938 &o->nb_stream_maps, o->nb_stream_maps + 1);
2939 m = &o->stream_maps[o->nb_stream_maps - 1];
2940
2941 m->file_index = file_idx;
2942 m->stream_index = i;
2943
2944 if (sync_file_idx >= 0) {
2945 m->sync_file_index = sync_file_idx;
2946 m->sync_stream_index = sync_stream_idx;
2947 } else {
2948 m->sync_file_index = file_idx;
2949 m->sync_stream_index = i;
2950 }
2951 }
2952
2953 if (!m) {
2954 av_log(NULL, AV_LOG_FATAL, "Stream map '%s' matches no streams.\n", arg);
2955 exit_program(1);
2956 }
2957
2958 av_freep(&map);
2959 return 0;
2960 }
2961
2962 static int opt_attach(OptionsContext *o, const char *opt, const char *arg)
2963 {
2964 o->attachments = grow_array(o->attachments, sizeof(*o->attachments),
2965 &o->nb_attachments, o->nb_attachments + 1);
2966 o->attachments[o->nb_attachments - 1] = arg;
2967 return 0;
2968 }
2969
2970 /**
2971 * Parse a metadata specifier in arg.
2972 * @param type metadata type is written here -- g(lobal)/s(tream)/c(hapter)/p(rogram)
2973 * @param index for type c/p, chapter/program index is written here
2974 * @param stream_spec for type s, the stream specifier is written here
2975 */
2976 static void parse_meta_type(char *arg, char *type, int *index, const char **stream_spec)
2977 {
2978 if (*arg) {
2979 *type = *arg;
2980 switch (*arg) {
2981 case 'g':
2982 break;
2983 case 's':
2984 if (*(++arg) && *arg != ':') {
2985 av_log(NULL, AV_LOG_FATAL, "Invalid metadata specifier %s.\n", arg);
2986 exit_program(1);
2987 }
2988 *stream_spec = *arg == ':' ? arg + 1 : "";
2989 break;
2990 case 'c':
2991 case 'p':
2992 if (*(++arg) == ':')
2993 *index = strtol(++arg, NULL, 0);
2994 break;
2995 default:
2996 av_log(NULL, AV_LOG_FATAL, "Invalid metadata type %c.\n", *arg);
2997 exit_program(1);
2998 }
2999 } else
3000 *type = 'g';
3001 }
3002
3003 static int copy_metadata(char *outspec, char *inspec, AVFormatContext *oc, AVFormatContext *ic, OptionsContext *o)
3004 {
3005 AVDictionary **meta_in = NULL;
3006 AVDictionary **meta_out;
3007 int i, ret = 0;
3008 char type_in, type_out;
3009 const char *istream_spec = NULL, *ostream_spec = NULL;
3010 int idx_in = 0, idx_out = 0;
3011
3012 parse_meta_type(inspec, &type_in, &idx_in, &istream_spec);
3013 parse_meta_type(outspec, &type_out, &idx_out, &ostream_spec);
3014
3015 if (type_in == 'g' || type_out == 'g')
3016 o->metadata_global_manual = 1;
3017 if (type_in == 's' || type_out == 's')
3018 o->metadata_streams_manual = 1;
3019 if (type_in == 'c' || type_out == 'c')
3020 o->metadata_chapters_manual = 1;
3021
3022 #define METADATA_CHECK_INDEX(index, nb_elems, desc)\
3023 if ((index) < 0 || (index) >= (nb_elems)) {\
3024 av_log(NULL, AV_LOG_FATAL, "Invalid %s index %d while processing metadata maps.\n",\
3025 (desc), (index));\
3026 exit_program(1);\
3027 }
3028
3029 #define SET_DICT(type, meta, context, index)\
3030 switch (type) {\
3031 case 'g':\
3032 meta = &context->metadata;\
3033 break;\
3034 case 'c':\
3035 METADATA_CHECK_INDEX(index, context->nb_chapters, "chapter")\
3036 meta = &context->chapters[index]->metadata;\
3037 break;\
3038 case 'p':\
3039 METADATA_CHECK_INDEX(index, context->nb_programs, "program")\
3040 meta = &context->programs[index]->metadata;\
3041 break;\
3042 }\
3043
3044 SET_DICT(type_in, meta_in, ic, idx_in);
3045 SET_DICT(type_out, meta_out, oc, idx_out);
3046
3047 /* for input streams choose first matching stream */
3048 if (type_in == 's') {
3049 for (i = 0; i < ic->nb_streams; i++) {
3050 if ((ret = check_stream_specifier(ic, ic->streams[i], istream_spec)) > 0) {
3051 meta_in = &ic->streams[i]->metadata;
3052 break;
3053 } else if (ret < 0)
3054 exit_program(1);
3055 }
3056 if (!meta_in) {
3057 av_log(NULL, AV_LOG_FATAL, "Stream specifier %s does not match any streams.\n", istream_spec);
3058 exit_program(1);
3059 }
3060 }
3061
3062 if (type_out == 's') {
3063 for (i = 0; i < oc->nb_streams; i++) {
3064 if ((ret = check_stream_specifier(oc, oc->streams[i], ostream_spec)) > 0) {
3065 meta_out = &oc->streams[i]->metadata;
3066 av_dict_copy(meta_out, *meta_in, AV_DICT_DONT_OVERWRITE);
3067 } else if (ret < 0)
3068 exit_program(1);
3069 }
3070 } else
3071 av_dict_copy(meta_out, *meta_in, AV_DICT_DONT_OVERWRITE);
3072
3073 return 0;
3074 }
3075
3076 static AVCodec *find_codec_or_die(const char *name, enum AVMediaType type, int encoder)
3077 {
3078 const char *codec_string = encoder ? "encoder" : "decoder";
3079 AVCodec *codec;
3080
3081 codec = encoder ?
3082 avcodec_find_encoder_by_name(name) :
3083 avcodec_find_decoder_by_name(name);
3084 if (!codec) {
3085 av_log(NULL, AV_LOG_FATAL, "Unknown %s '%s'\n", codec_string, name);
3086 exit_program(1);
3087 }
3088 if (codec->type != type) {
3089 av_log(NULL, AV_LOG_FATAL, "Invalid %s type '%s'\n", codec_string, name);
3090 exit_program(1);
3091 }
3092 return codec;
3093 }
3094
3095 static AVCodec *choose_decoder(OptionsContext *o, AVFormatContext *s, AVStream *st)
3096 {
3097 char *codec_name = NULL;
3098
3099 MATCH_PER_STREAM_OPT(codec_names, str, codec_name, s, st);
3100 if (codec_name) {
3101 AVCodec *codec = find_codec_or_die(codec_name, st->codec->codec_type, 0);
3102 st->codec->codec_id = codec->id;
3103 return codec;
3104 } else
3105 return avcodec_find_decoder(st->codec->codec_id);
3106 }
3107
3108 /**
3109 * Add all the streams from the given input file to the global
3110 * list of input streams.
3111 */
3112 static void add_input_streams(OptionsContext *o, AVFormatContext *ic)
3113 {
3114 int i;
3115
3116 for (i = 0; i < ic->nb_streams; i++) {
3117 AVStream *st = ic->streams[i];
3118 AVCodecContext *dec = st->codec;
3119 InputStream *ist;
3120
3121 input_streams = grow_array(input_streams, sizeof(*input_streams), &nb_input_streams, nb_input_streams + 1);
3122 ist = &input_streams[nb_input_streams - 1];
3123 ist->st = st;
3124 ist->file_index = nb_input_files;
3125 ist->discard = 1;
3126 ist->opts = filter_codec_opts(codec_opts, ist->st->codec->codec_id, ic, st);
3127
3128 ist->ts_scale = 1.0;
3129 MATCH_PER_STREAM_OPT(ts_scale, dbl, ist->ts_scale, ic, st);
3130
3131 ist->dec = choose_decoder(o, ic, st);
3132
3133 switch (dec->codec_type) {
3134 case AVMEDIA_TYPE_AUDIO:
3135 if (o->audio_disable)
3136 st->discard = AVDISCARD_ALL;
3137 break;
3138 case AVMEDIA_TYPE_VIDEO:
3139 if (dec->lowres) {
3140 dec->flags |= CODEC_FLAG_EMU_EDGE;
3141 dec->height >>= dec->lowres;
3142 dec->width >>= dec->lowres;
3143 }
3144
3145 if (o->video_disable)
3146 st->discard = AVDISCARD_ALL;
3147 else if (video_discard)
3148 st->discard = video_discard;
3149 break;
3150 case AVMEDIA_TYPE_DATA:
3151 break;
3152 case AVMEDIA_TYPE_SUBTITLE:
3153 if (o->subtitle_disable)
3154 st->discard = AVDISCARD_ALL;
3155 break;
3156 case AVMEDIA_TYPE_ATTACHMENT:
3157 case AVMEDIA_TYPE_UNKNOWN:
3158 break;
3159 default:
3160 abort();
3161 }
3162 }
3163 }
3164
3165 static void assert_file_overwrite(const char *filename)
3166 {
3167 if (!file_overwrite &&
3168 (strchr(filename, ':') == NULL || filename[1] == ':' ||
3169 av_strstart(filename, "file:", NULL))) {
3170 if (avio_check(filename, 0) == 0) {
3171 if (!using_stdin) {
3172 fprintf(stderr,"File '%s' already exists. Overwrite ? [y/N] ", filename);
3173 fflush(stderr);
3174 if (!read_yesno()) {
3175 fprintf(stderr, "Not overwriting - exiting\n");
3176 exit_program(1);
3177 }
3178 }
3179 else {
3180 fprintf(stderr,"File '%s' already exists. Exiting.\n", filename);
3181 exit_program(1);
3182 }
3183 }
3184 }
3185 }
3186
3187 static void dump_attachment(AVStream *st, const char *filename)
3188 {
3189 int ret;
3190 AVIOContext *out = NULL;
3191 AVDictionaryEntry *e;
3192
3193 if (!st->codec->extradata_size) {
3194 av_log(NULL, AV_LOG_WARNING, "No extradata to dump in stream #%d:%d.\n",
3195 nb_input_files - 1, st->index);
3196 return;
3197 }
3198 if (!*filename && (e = av_dict_get(st->metadata, "filename", NULL, 0)))
3199 filename = e->value;
3200 if (!*filename) {
3201 av_log(NULL, AV_LOG_FATAL, "No filename specified and no 'filename' tag"
3202 "in stream #%d:%d.\n", nb_input_files - 1, st->index);
3203 exit_program(1);
3204 }
3205
3206 assert_file_overwrite(filename);
3207
3208 if ((ret = avio_open2(&out, filename, AVIO_FLAG_WRITE, &int_cb, NULL)) < 0) {
3209 av_log(NULL, AV_LOG_FATAL, "Could not open file %s for writing.\n