Replace "-vfilters" by "-vf" in regtests. Should fix regtest breakage.
[libav.git] / ffplay.c
1 /*
2 * FFplay : Simple Media Player based on the FFmpeg libraries
3 * Copyright (c) 2003 Fabrice Bellard
4 *
5 * This file is part of FFmpeg.
6 *
7 * FFmpeg is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
11 *
12 * FFmpeg is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with FFmpeg; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20 */
21
22 #include "config.h"
23 #include <inttypes.h>
24 #include <math.h>
25 #include <limits.h>
26 #include "libavutil/avstring.h"
27 #include "libavutil/pixdesc.h"
28 #include "libavformat/avformat.h"
29 #include "libavdevice/avdevice.h"
30 #include "libswscale/swscale.h"
31 #include "libavcodec/audioconvert.h"
32 #include "libavcodec/colorspace.h"
33 #include "libavcodec/opt.h"
34 #include "libavcodec/avfft.h"
35
36 #if CONFIG_AVFILTER
37 # include "libavfilter/avfilter.h"
38 # include "libavfilter/avfiltergraph.h"
39 # include "libavfilter/graphparser.h"
40 #endif
41
42 #include "cmdutils.h"
43
44 #include <SDL.h>
45 #include <SDL_thread.h>
46
47 #ifdef __MINGW32__
48 #undef main /* We don't want SDL to override our main() */
49 #endif
50
51 #include <unistd.h>
52 #include <assert.h>
53
54 const char program_name[] = "FFplay";
55 const int program_birth_year = 2003;
56
57 //#define DEBUG_SYNC
58
59 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
60 #define MIN_AUDIOQ_SIZE (20 * 16 * 1024)
61 #define MIN_FRAMES 5
62
63 /* SDL audio buffer size, in samples. Should be small to have precise
64 A/V sync as SDL does not have hardware buffer fullness info. */
65 #define SDL_AUDIO_BUFFER_SIZE 1024
66
67 /* no AV sync correction is done if below the AV sync threshold */
68 #define AV_SYNC_THRESHOLD 0.01
69 /* no AV correction is done if too big error */
70 #define AV_NOSYNC_THRESHOLD 10.0
71
72 #define FRAME_SKIP_FACTOR 0.05
73
74 /* maximum audio speed change to get correct sync */
75 #define SAMPLE_CORRECTION_PERCENT_MAX 10
76
77 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
78 #define AUDIO_DIFF_AVG_NB 20
79
80 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
81 #define SAMPLE_ARRAY_SIZE (2*65536)
82
83 static int sws_flags = SWS_BICUBIC;
84
85 typedef struct PacketQueue {
86 AVPacketList *first_pkt, *last_pkt;
87 int nb_packets;
88 int size;
89 int abort_request;
90 SDL_mutex *mutex;
91 SDL_cond *cond;
92 } PacketQueue;
93
94 #define VIDEO_PICTURE_QUEUE_SIZE 2
95 #define SUBPICTURE_QUEUE_SIZE 4
96
97 typedef struct VideoPicture {
98 double pts; ///<presentation time stamp for this picture
99 double target_clock; ///<av_gettime() time at which this should be displayed ideally
100 int64_t pos; ///<byte position in file
101 SDL_Overlay *bmp;
102 int width, height; /* source height & width */
103 int allocated;
104 enum PixelFormat pix_fmt;
105
106 #if CONFIG_AVFILTER
107 AVFilterPicRef *picref;
108 #endif
109 } VideoPicture;
110
111 typedef struct SubPicture {
112 double pts; /* presentation time stamp for this picture */
113 AVSubtitle sub;
114 } SubPicture;
115
116 enum {
117 AV_SYNC_AUDIO_MASTER, /* default choice */
118 AV_SYNC_VIDEO_MASTER,
119 AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
120 };
121
122 typedef struct VideoState {
123 SDL_Thread *parse_tid;
124 SDL_Thread *video_tid;
125 SDL_Thread *refresh_tid;
126 AVInputFormat *iformat;
127 int no_background;
128 int abort_request;
129 int paused;
130 int last_paused;
131 int seek_req;
132 int seek_flags;
133 int64_t seek_pos;
134 int64_t seek_rel;
135 int read_pause_return;
136 AVFormatContext *ic;
137 int dtg_active_format;
138
139 int audio_stream;
140
141 int av_sync_type;
142 double external_clock; /* external clock base */
143 int64_t external_clock_time;
144
145 double audio_clock;
146 double audio_diff_cum; /* used for AV difference average computation */
147 double audio_diff_avg_coef;
148 double audio_diff_threshold;
149 int audio_diff_avg_count;
150 AVStream *audio_st;
151 PacketQueue audioq;
152 int audio_hw_buf_size;
153 /* samples output by the codec. we reserve more space for avsync
154 compensation */
155 DECLARE_ALIGNED(16,uint8_t,audio_buf1)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
156 DECLARE_ALIGNED(16,uint8_t,audio_buf2)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
157 uint8_t *audio_buf;
158 unsigned int audio_buf_size; /* in bytes */
159 int audio_buf_index; /* in bytes */
160 AVPacket audio_pkt_temp;
161 AVPacket audio_pkt;
162 enum SampleFormat audio_src_fmt;
163 AVAudioConvert *reformat_ctx;
164
165 int show_audio; /* if true, display audio samples */
166 int16_t sample_array[SAMPLE_ARRAY_SIZE];
167 int sample_array_index;
168 int last_i_start;
169 RDFTContext *rdft;
170 int rdft_bits;
171 int xpos;
172
173 SDL_Thread *subtitle_tid;
174 int subtitle_stream;
175 int subtitle_stream_changed;
176 AVStream *subtitle_st;
177 PacketQueue subtitleq;
178 SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
179 int subpq_size, subpq_rindex, subpq_windex;
180 SDL_mutex *subpq_mutex;
181 SDL_cond *subpq_cond;
182
183 double frame_timer;
184 double frame_last_pts;
185 double frame_last_delay;
186 double video_clock; ///<pts of last decoded frame / predicted pts of next decoded frame
187 int video_stream;
188 AVStream *video_st;
189 PacketQueue videoq;
190 double video_current_pts; ///<current displayed pts (different from video_clock if frame fifos are used)
191 double video_current_pts_drift; ///<video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
192 int64_t video_current_pos; ///<current displayed file pos
193 VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
194 int pictq_size, pictq_rindex, pictq_windex;
195 SDL_mutex *pictq_mutex;
196 SDL_cond *pictq_cond;
197 #if !CONFIG_AVFILTER
198 struct SwsContext *img_convert_ctx;
199 #endif
200
201 // QETimer *video_timer;
202 char filename[1024];
203 int width, height, xleft, ytop;
204
205 int64_t faulty_pts;
206 int64_t faulty_dts;
207 int64_t last_dts_for_fault_detection;
208 int64_t last_pts_for_fault_detection;
209
210 #if CONFIG_AVFILTER
211 AVFilterContext *out_video_filter; ///<the last filter in the video chain
212 #endif
213
214 float skip_frames;
215 float skip_frames_index;
216 int refresh;
217 } VideoState;
218
219 static void show_help(void);
220 static int audio_write_get_buf_size(VideoState *is);
221
222 /* options specified by the user */
223 static AVInputFormat *file_iformat;
224 static const char *input_filename;
225 static const char *window_title;
226 static int fs_screen_width;
227 static int fs_screen_height;
228 static int screen_width = 0;
229 static int screen_height = 0;
230 static int frame_width = 0;
231 static int frame_height = 0;
232 static enum PixelFormat frame_pix_fmt = PIX_FMT_NONE;
233 static int audio_disable;
234 static int video_disable;
235 static int wanted_stream[AVMEDIA_TYPE_NB]={
236 [AVMEDIA_TYPE_AUDIO]=-1,
237 [AVMEDIA_TYPE_VIDEO]=-1,
238 [AVMEDIA_TYPE_SUBTITLE]=-1,
239 };
240 static int seek_by_bytes=-1;
241 static int display_disable;
242 static int show_status = 1;
243 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
244 static int64_t start_time = AV_NOPTS_VALUE;
245 static int64_t duration = AV_NOPTS_VALUE;
246 static int debug = 0;
247 static int debug_mv = 0;
248 static int step = 0;
249 static int thread_count = 1;
250 static int workaround_bugs = 1;
251 static int fast = 0;
252 static int genpts = 0;
253 static int lowres = 0;
254 static int idct = FF_IDCT_AUTO;
255 static enum AVDiscard skip_frame= AVDISCARD_DEFAULT;
256 static enum AVDiscard skip_idct= AVDISCARD_DEFAULT;
257 static enum AVDiscard skip_loop_filter= AVDISCARD_DEFAULT;
258 static int error_recognition = FF_ER_CAREFUL;
259 static int error_concealment = 3;
260 static int decoder_reorder_pts= -1;
261 static int autoexit;
262 static int loop=1;
263 static int framedrop=1;
264
265 static int rdftspeed=20;
266 #if CONFIG_AVFILTER
267 static char *vfilters = NULL;
268 #endif
269
270 /* current context */
271 static int is_full_screen;
272 static VideoState *cur_stream;
273 static int64_t audio_callback_time;
274
275 static AVPacket flush_pkt;
276
277 #define FF_ALLOC_EVENT (SDL_USEREVENT)
278 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
279 #define FF_QUIT_EVENT (SDL_USEREVENT + 2)
280
281 static SDL_Surface *screen;
282
283 static int packet_queue_put(PacketQueue *q, AVPacket *pkt);
284
285 /* packet queue handling */
286 static void packet_queue_init(PacketQueue *q)
287 {
288 memset(q, 0, sizeof(PacketQueue));
289 q->mutex = SDL_CreateMutex();
290 q->cond = SDL_CreateCond();
291 packet_queue_put(q, &flush_pkt);
292 }
293
294 static void packet_queue_flush(PacketQueue *q)
295 {
296 AVPacketList *pkt, *pkt1;
297
298 SDL_LockMutex(q->mutex);
299 for(pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
300 pkt1 = pkt->next;
301 av_free_packet(&pkt->pkt);
302 av_freep(&pkt);
303 }
304 q->last_pkt = NULL;
305 q->first_pkt = NULL;
306 q->nb_packets = 0;
307 q->size = 0;
308 SDL_UnlockMutex(q->mutex);
309 }
310
311 static void packet_queue_end(PacketQueue *q)
312 {
313 packet_queue_flush(q);
314 SDL_DestroyMutex(q->mutex);
315 SDL_DestroyCond(q->cond);
316 }
317
318 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
319 {
320 AVPacketList *pkt1;
321
322 /* duplicate the packet */
323 if (pkt!=&flush_pkt && av_dup_packet(pkt) < 0)
324 return -1;
325
326 pkt1 = av_malloc(sizeof(AVPacketList));
327 if (!pkt1)
328 return -1;
329 pkt1->pkt = *pkt;
330 pkt1->next = NULL;
331
332
333 SDL_LockMutex(q->mutex);
334
335 if (!q->last_pkt)
336
337 q->first_pkt = pkt1;
338 else
339 q->last_pkt->next = pkt1;
340 q->last_pkt = pkt1;
341 q->nb_packets++;
342 q->size += pkt1->pkt.size + sizeof(*pkt1);
343 /* XXX: should duplicate packet data in DV case */
344 SDL_CondSignal(q->cond);
345
346 SDL_UnlockMutex(q->mutex);
347 return 0;
348 }
349
350 static void packet_queue_abort(PacketQueue *q)
351 {
352 SDL_LockMutex(q->mutex);
353
354 q->abort_request = 1;
355
356 SDL_CondSignal(q->cond);
357
358 SDL_UnlockMutex(q->mutex);
359 }
360
361 /* return < 0 if aborted, 0 if no packet and > 0 if packet. */
362 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
363 {
364 AVPacketList *pkt1;
365 int ret;
366
367 SDL_LockMutex(q->mutex);
368
369 for(;;) {
370 if (q->abort_request) {
371 ret = -1;
372 break;
373 }
374
375 pkt1 = q->first_pkt;
376 if (pkt1) {
377 q->first_pkt = pkt1->next;
378 if (!q->first_pkt)
379 q->last_pkt = NULL;
380 q->nb_packets--;
381 q->size -= pkt1->pkt.size + sizeof(*pkt1);
382 *pkt = pkt1->pkt;
383 av_free(pkt1);
384 ret = 1;
385 break;
386 } else if (!block) {
387 ret = 0;
388 break;
389 } else {
390 SDL_CondWait(q->cond, q->mutex);
391 }
392 }
393 SDL_UnlockMutex(q->mutex);
394 return ret;
395 }
396
397 static inline void fill_rectangle(SDL_Surface *screen,
398 int x, int y, int w, int h, int color)
399 {
400 SDL_Rect rect;
401 rect.x = x;
402 rect.y = y;
403 rect.w = w;
404 rect.h = h;
405 SDL_FillRect(screen, &rect, color);
406 }
407
408 #if 0
409 /* draw only the border of a rectangle */
410 void fill_border(VideoState *s, int x, int y, int w, int h, int color)
411 {
412 int w1, w2, h1, h2;
413
414 /* fill the background */
415 w1 = x;
416 if (w1 < 0)
417 w1 = 0;
418 w2 = s->width - (x + w);
419 if (w2 < 0)
420 w2 = 0;
421 h1 = y;
422 if (h1 < 0)
423 h1 = 0;
424 h2 = s->height - (y + h);
425 if (h2 < 0)
426 h2 = 0;
427 fill_rectangle(screen,
428 s->xleft, s->ytop,
429 w1, s->height,
430 color);
431 fill_rectangle(screen,
432 s->xleft + s->width - w2, s->ytop,
433 w2, s->height,
434 color);
435 fill_rectangle(screen,
436 s->xleft + w1, s->ytop,
437 s->width - w1 - w2, h1,
438 color);
439 fill_rectangle(screen,
440 s->xleft + w1, s->ytop + s->height - h2,
441 s->width - w1 - w2, h2,
442 color);
443 }
444 #endif
445
446 #define ALPHA_BLEND(a, oldp, newp, s)\
447 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
448
449 #define RGBA_IN(r, g, b, a, s)\
450 {\
451 unsigned int v = ((const uint32_t *)(s))[0];\
452 a = (v >> 24) & 0xff;\
453 r = (v >> 16) & 0xff;\
454 g = (v >> 8) & 0xff;\
455 b = v & 0xff;\
456 }
457
458 #define YUVA_IN(y, u, v, a, s, pal)\
459 {\
460 unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
461 a = (val >> 24) & 0xff;\
462 y = (val >> 16) & 0xff;\
463 u = (val >> 8) & 0xff;\
464 v = val & 0xff;\
465 }
466
467 #define YUVA_OUT(d, y, u, v, a)\
468 {\
469 ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
470 }
471
472
473 #define BPP 1
474
475 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
476 {
477 int wrap, wrap3, width2, skip2;
478 int y, u, v, a, u1, v1, a1, w, h;
479 uint8_t *lum, *cb, *cr;
480 const uint8_t *p;
481 const uint32_t *pal;
482 int dstx, dsty, dstw, dsth;
483
484 dstw = av_clip(rect->w, 0, imgw);
485 dsth = av_clip(rect->h, 0, imgh);
486 dstx = av_clip(rect->x, 0, imgw - dstw);
487 dsty = av_clip(rect->y, 0, imgh - dsth);
488 lum = dst->data[0] + dsty * dst->linesize[0];
489 cb = dst->data[1] + (dsty >> 1) * dst->linesize[1];
490 cr = dst->data[2] + (dsty >> 1) * dst->linesize[2];
491
492 width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
493 skip2 = dstx >> 1;
494 wrap = dst->linesize[0];
495 wrap3 = rect->pict.linesize[0];
496 p = rect->pict.data[0];
497 pal = (const uint32_t *)rect->pict.data[1]; /* Now in YCrCb! */
498
499 if (dsty & 1) {
500 lum += dstx;
501 cb += skip2;
502 cr += skip2;
503
504 if (dstx & 1) {
505 YUVA_IN(y, u, v, a, p, pal);
506 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
507 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
508 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
509 cb++;
510 cr++;
511 lum++;
512 p += BPP;
513 }
514 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
515 YUVA_IN(y, u, v, a, p, pal);
516 u1 = u;
517 v1 = v;
518 a1 = a;
519 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
520
521 YUVA_IN(y, u, v, a, p + BPP, pal);
522 u1 += u;
523 v1 += v;
524 a1 += a;
525 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
526 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
527 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
528 cb++;
529 cr++;
530 p += 2 * BPP;
531 lum += 2;
532 }
533 if (w) {
534 YUVA_IN(y, u, v, a, p, pal);
535 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
536 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
537 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
538 p++;
539 lum++;
540 }
541 p += wrap3 - dstw * BPP;
542 lum += wrap - dstw - dstx;
543 cb += dst->linesize[1] - width2 - skip2;
544 cr += dst->linesize[2] - width2 - skip2;
545 }
546 for(h = dsth - (dsty & 1); h >= 2; h -= 2) {
547 lum += dstx;
548 cb += skip2;
549 cr += skip2;
550
551 if (dstx & 1) {
552 YUVA_IN(y, u, v, a, p, pal);
553 u1 = u;
554 v1 = v;
555 a1 = a;
556 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
557 p += wrap3;
558 lum += wrap;
559 YUVA_IN(y, u, v, a, p, pal);
560 u1 += u;
561 v1 += v;
562 a1 += a;
563 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
564 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
565 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
566 cb++;
567 cr++;
568 p += -wrap3 + BPP;
569 lum += -wrap + 1;
570 }
571 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
572 YUVA_IN(y, u, v, a, p, pal);
573 u1 = u;
574 v1 = v;
575 a1 = a;
576 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
577
578 YUVA_IN(y, u, v, a, p + BPP, pal);
579 u1 += u;
580 v1 += v;
581 a1 += a;
582 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
583 p += wrap3;
584 lum += wrap;
585
586 YUVA_IN(y, u, v, a, p, pal);
587 u1 += u;
588 v1 += v;
589 a1 += a;
590 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
591
592 YUVA_IN(y, u, v, a, p + BPP, pal);
593 u1 += u;
594 v1 += v;
595 a1 += a;
596 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
597
598 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
599 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
600
601 cb++;
602 cr++;
603 p += -wrap3 + 2 * BPP;
604 lum += -wrap + 2;
605 }
606 if (w) {
607 YUVA_IN(y, u, v, a, p, pal);
608 u1 = u;
609 v1 = v;
610 a1 = a;
611 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
612 p += wrap3;
613 lum += wrap;
614 YUVA_IN(y, u, v, a, p, pal);
615 u1 += u;
616 v1 += v;
617 a1 += a;
618 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
619 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
620 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
621 cb++;
622 cr++;
623 p += -wrap3 + BPP;
624 lum += -wrap + 1;
625 }
626 p += wrap3 + (wrap3 - dstw * BPP);
627 lum += wrap + (wrap - dstw - dstx);
628 cb += dst->linesize[1] - width2 - skip2;
629 cr += dst->linesize[2] - width2 - skip2;
630 }
631 /* handle odd height */
632 if (h) {
633 lum += dstx;
634 cb += skip2;
635 cr += skip2;
636
637 if (dstx & 1) {
638 YUVA_IN(y, u, v, a, p, pal);
639 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
640 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
641 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
642 cb++;
643 cr++;
644 lum++;
645 p += BPP;
646 }
647 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
648 YUVA_IN(y, u, v, a, p, pal);
649 u1 = u;
650 v1 = v;
651 a1 = a;
652 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
653
654 YUVA_IN(y, u, v, a, p + BPP, pal);
655 u1 += u;
656 v1 += v;
657 a1 += a;
658 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
659 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
660 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
661 cb++;
662 cr++;
663 p += 2 * BPP;
664 lum += 2;
665 }
666 if (w) {
667 YUVA_IN(y, u, v, a, p, pal);
668 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
669 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
670 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
671 }
672 }
673 }
674
675 static void free_subpicture(SubPicture *sp)
676 {
677 int i;
678
679 for (i = 0; i < sp->sub.num_rects; i++)
680 {
681 av_freep(&sp->sub.rects[i]->pict.data[0]);
682 av_freep(&sp->sub.rects[i]->pict.data[1]);
683 av_freep(&sp->sub.rects[i]);
684 }
685
686 av_free(sp->sub.rects);
687
688 memset(&sp->sub, 0, sizeof(AVSubtitle));
689 }
690
691 static void video_image_display(VideoState *is)
692 {
693 VideoPicture *vp;
694 SubPicture *sp;
695 AVPicture pict;
696 float aspect_ratio;
697 int width, height, x, y;
698 SDL_Rect rect;
699 int i;
700
701 vp = &is->pictq[is->pictq_rindex];
702 if (vp->bmp) {
703 #if CONFIG_AVFILTER
704 if (vp->picref->pixel_aspect.num == 0)
705 aspect_ratio = 0;
706 else
707 aspect_ratio = av_q2d(vp->picref->pixel_aspect);
708 #else
709
710 /* XXX: use variable in the frame */
711 if (is->video_st->sample_aspect_ratio.num)
712 aspect_ratio = av_q2d(is->video_st->sample_aspect_ratio);
713 else if (is->video_st->codec->sample_aspect_ratio.num)
714 aspect_ratio = av_q2d(is->video_st->codec->sample_aspect_ratio);
715 else
716 aspect_ratio = 0;
717 #endif
718 if (aspect_ratio <= 0.0)
719 aspect_ratio = 1.0;
720 aspect_ratio *= (float)vp->width / (float)vp->height;
721 /* if an active format is indicated, then it overrides the
722 mpeg format */
723 #if 0
724 if (is->video_st->codec->dtg_active_format != is->dtg_active_format) {
725 is->dtg_active_format = is->video_st->codec->dtg_active_format;
726 printf("dtg_active_format=%d\n", is->dtg_active_format);
727 }
728 #endif
729 #if 0
730 switch(is->video_st->codec->dtg_active_format) {
731 case FF_DTG_AFD_SAME:
732 default:
733 /* nothing to do */
734 break;
735 case FF_DTG_AFD_4_3:
736 aspect_ratio = 4.0 / 3.0;
737 break;
738 case FF_DTG_AFD_16_9:
739 aspect_ratio = 16.0 / 9.0;
740 break;
741 case FF_DTG_AFD_14_9:
742 aspect_ratio = 14.0 / 9.0;
743 break;
744 case FF_DTG_AFD_4_3_SP_14_9:
745 aspect_ratio = 14.0 / 9.0;
746 break;
747 case FF_DTG_AFD_16_9_SP_14_9:
748 aspect_ratio = 14.0 / 9.0;
749 break;
750 case FF_DTG_AFD_SP_4_3:
751 aspect_ratio = 4.0 / 3.0;
752 break;
753 }
754 #endif
755
756 if (is->subtitle_st)
757 {
758 if (is->subpq_size > 0)
759 {
760 sp = &is->subpq[is->subpq_rindex];
761
762 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000))
763 {
764 SDL_LockYUVOverlay (vp->bmp);
765
766 pict.data[0] = vp->bmp->pixels[0];
767 pict.data[1] = vp->bmp->pixels[2];
768 pict.data[2] = vp->bmp->pixels[1];
769
770 pict.linesize[0] = vp->bmp->pitches[0];
771 pict.linesize[1] = vp->bmp->pitches[2];
772 pict.linesize[2] = vp->bmp->pitches[1];
773
774 for (i = 0; i < sp->sub.num_rects; i++)
775 blend_subrect(&pict, sp->sub.rects[i],
776 vp->bmp->w, vp->bmp->h);
777
778 SDL_UnlockYUVOverlay (vp->bmp);
779 }
780 }
781 }
782
783
784 /* XXX: we suppose the screen has a 1.0 pixel ratio */
785 height = is->height;
786 width = ((int)rint(height * aspect_ratio)) & ~1;
787 if (width > is->width) {
788 width = is->width;
789 height = ((int)rint(width / aspect_ratio)) & ~1;
790 }
791 x = (is->width - width) / 2;
792 y = (is->height - height) / 2;
793 if (!is->no_background) {
794 /* fill the background */
795 // fill_border(is, x, y, width, height, QERGB(0x00, 0x00, 0x00));
796 } else {
797 is->no_background = 0;
798 }
799 rect.x = is->xleft + x;
800 rect.y = is->ytop + y;
801 rect.w = width;
802 rect.h = height;
803 SDL_DisplayYUVOverlay(vp->bmp, &rect);
804 } else {
805 #if 0
806 fill_rectangle(screen,
807 is->xleft, is->ytop, is->width, is->height,
808 QERGB(0x00, 0x00, 0x00));
809 #endif
810 }
811 }
812
813 static inline int compute_mod(int a, int b)
814 {
815 a = a % b;
816 if (a >= 0)
817 return a;
818 else
819 return a + b;
820 }
821
822 static void video_audio_display(VideoState *s)
823 {
824 int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
825 int ch, channels, h, h2, bgcolor, fgcolor;
826 int16_t time_diff;
827 int rdft_bits, nb_freq;
828
829 for(rdft_bits=1; (1<<rdft_bits)<2*s->height; rdft_bits++)
830 ;
831 nb_freq= 1<<(rdft_bits-1);
832
833 /* compute display index : center on currently output samples */
834 channels = s->audio_st->codec->channels;
835 nb_display_channels = channels;
836 if (!s->paused) {
837 int data_used= s->show_audio==1 ? s->width : (2*nb_freq);
838 n = 2 * channels;
839 delay = audio_write_get_buf_size(s);
840 delay /= n;
841
842 /* to be more precise, we take into account the time spent since
843 the last buffer computation */
844 if (audio_callback_time) {
845 time_diff = av_gettime() - audio_callback_time;
846 delay -= (time_diff * s->audio_st->codec->sample_rate) / 1000000;
847 }
848
849 delay += 2*data_used;
850 if (delay < data_used)
851 delay = data_used;
852
853 i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
854 if(s->show_audio==1){
855 h= INT_MIN;
856 for(i=0; i<1000; i+=channels){
857 int idx= (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
858 int a= s->sample_array[idx];
859 int b= s->sample_array[(idx + 4*channels)%SAMPLE_ARRAY_SIZE];
860 int c= s->sample_array[(idx + 5*channels)%SAMPLE_ARRAY_SIZE];
861 int d= s->sample_array[(idx + 9*channels)%SAMPLE_ARRAY_SIZE];
862 int score= a-d;
863 if(h<score && (b^c)<0){
864 h= score;
865 i_start= idx;
866 }
867 }
868 }
869
870 s->last_i_start = i_start;
871 } else {
872 i_start = s->last_i_start;
873 }
874
875 bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
876 if(s->show_audio==1){
877 fill_rectangle(screen,
878 s->xleft, s->ytop, s->width, s->height,
879 bgcolor);
880
881 fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
882
883 /* total height for one channel */
884 h = s->height / nb_display_channels;
885 /* graph height / 2 */
886 h2 = (h * 9) / 20;
887 for(ch = 0;ch < nb_display_channels; ch++) {
888 i = i_start + ch;
889 y1 = s->ytop + ch * h + (h / 2); /* position of center line */
890 for(x = 0; x < s->width; x++) {
891 y = (s->sample_array[i] * h2) >> 15;
892 if (y < 0) {
893 y = -y;
894 ys = y1 - y;
895 } else {
896 ys = y1;
897 }
898 fill_rectangle(screen,
899 s->xleft + x, ys, 1, y,
900 fgcolor);
901 i += channels;
902 if (i >= SAMPLE_ARRAY_SIZE)
903 i -= SAMPLE_ARRAY_SIZE;
904 }
905 }
906
907 fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
908
909 for(ch = 1;ch < nb_display_channels; ch++) {
910 y = s->ytop + ch * h;
911 fill_rectangle(screen,
912 s->xleft, y, s->width, 1,
913 fgcolor);
914 }
915 SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
916 }else{
917 nb_display_channels= FFMIN(nb_display_channels, 2);
918 if(rdft_bits != s->rdft_bits){
919 av_rdft_end(s->rdft);
920 s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
921 s->rdft_bits= rdft_bits;
922 }
923 {
924 FFTSample data[2][2*nb_freq];
925 for(ch = 0;ch < nb_display_channels; ch++) {
926 i = i_start + ch;
927 for(x = 0; x < 2*nb_freq; x++) {
928 double w= (x-nb_freq)*(1.0/nb_freq);
929 data[ch][x]= s->sample_array[i]*(1.0-w*w);
930 i += channels;
931 if (i >= SAMPLE_ARRAY_SIZE)
932 i -= SAMPLE_ARRAY_SIZE;
933 }
934 av_rdft_calc(s->rdft, data[ch]);
935 }
936 //least efficient way to do this, we should of course directly access it but its more than fast enough
937 for(y=0; y<s->height; y++){
938 double w= 1/sqrt(nb_freq);
939 int a= sqrt(w*sqrt(data[0][2*y+0]*data[0][2*y+0] + data[0][2*y+1]*data[0][2*y+1]));
940 int b= sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0] + data[1][2*y+1]*data[1][2*y+1]));
941 a= FFMIN(a,255);
942 b= FFMIN(b,255);
943 fgcolor = SDL_MapRGB(screen->format, a, b, (a+b)/2);
944
945 fill_rectangle(screen,
946 s->xpos, s->height-y, 1, 1,
947 fgcolor);
948 }
949 }
950 SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
951 s->xpos++;
952 if(s->xpos >= s->width)
953 s->xpos= s->xleft;
954 }
955 }
956
957 static int video_open(VideoState *is){
958 int flags = SDL_HWSURFACE|SDL_ASYNCBLIT|SDL_HWACCEL;
959 int w,h;
960
961 if(is_full_screen) flags |= SDL_FULLSCREEN;
962 else flags |= SDL_RESIZABLE;
963
964 if (is_full_screen && fs_screen_width) {
965 w = fs_screen_width;
966 h = fs_screen_height;
967 } else if(!is_full_screen && screen_width){
968 w = screen_width;
969 h = screen_height;
970 #if CONFIG_AVFILTER
971 }else if (is->out_video_filter && is->out_video_filter->inputs[0]){
972 w = is->out_video_filter->inputs[0]->w;
973 h = is->out_video_filter->inputs[0]->h;
974 #else
975 }else if (is->video_st && is->video_st->codec->width){
976 w = is->video_st->codec->width;
977 h = is->video_st->codec->height;
978 #endif
979 } else {
980 w = 640;
981 h = 480;
982 }
983 if(screen && is->width == screen->w && screen->w == w
984 && is->height== screen->h && screen->h == h)
985 return 0;
986
987 #ifndef __APPLE__
988 screen = SDL_SetVideoMode(w, h, 0, flags);
989 #else
990 /* setting bits_per_pixel = 0 or 32 causes blank video on OS X */
991 screen = SDL_SetVideoMode(w, h, 24, flags);
992 #endif
993 if (!screen) {
994 fprintf(stderr, "SDL: could not set video mode - exiting\n");
995 return -1;
996 }
997 if (!window_title)
998 window_title = input_filename;
999 SDL_WM_SetCaption(window_title, window_title);
1000
1001 is->width = screen->w;
1002 is->height = screen->h;
1003
1004 return 0;
1005 }
1006
1007 /* display the current picture, if any */
1008 static void video_display(VideoState *is)
1009 {
1010 if(!screen)
1011 video_open(cur_stream);
1012 if (is->audio_st && is->show_audio)
1013 video_audio_display(is);
1014 else if (is->video_st)
1015 video_image_display(is);
1016 }
1017
1018 static int refresh_thread(void *opaque)
1019 {
1020 VideoState *is= opaque;
1021 while(!is->abort_request){
1022 SDL_Event event;
1023 event.type = FF_REFRESH_EVENT;
1024 event.user.data1 = opaque;
1025 if(!is->refresh){
1026 is->refresh=1;
1027 SDL_PushEvent(&event);
1028 }
1029 usleep(is->audio_st && is->show_audio ? rdftspeed*1000 : 5000); //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
1030 }
1031 return 0;
1032 }
1033
1034 /* get the current audio clock value */
1035 static double get_audio_clock(VideoState *is)
1036 {
1037 double pts;
1038 int hw_buf_size, bytes_per_sec;
1039 pts = is->audio_clock;
1040 hw_buf_size = audio_write_get_buf_size(is);
1041 bytes_per_sec = 0;
1042 if (is->audio_st) {
1043 bytes_per_sec = is->audio_st->codec->sample_rate *
1044 2 * is->audio_st->codec->channels;
1045 }
1046 if (bytes_per_sec)
1047 pts -= (double)hw_buf_size / bytes_per_sec;
1048 return pts;
1049 }
1050
1051 /* get the current video clock value */
1052 static double get_video_clock(VideoState *is)
1053 {
1054 if (is->paused) {
1055 return is->video_current_pts;
1056 } else {
1057 return is->video_current_pts_drift + av_gettime() / 1000000.0;
1058 }
1059 }
1060
1061 /* get the current external clock value */
1062 static double get_external_clock(VideoState *is)
1063 {
1064 int64_t ti;
1065 ti = av_gettime();
1066 return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1067 }
1068
1069 /* get the current master clock value */
1070 static double get_master_clock(VideoState *is)
1071 {
1072 double val;
1073
1074 if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1075 if (is->video_st)
1076 val = get_video_clock(is);
1077 else
1078 val = get_audio_clock(is);
1079 } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1080 if (is->audio_st)
1081 val = get_audio_clock(is);
1082 else
1083 val = get_video_clock(is);
1084 } else {
1085 val = get_external_clock(is);
1086 }
1087 return val;
1088 }
1089
1090 /* seek in the stream */
1091 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1092 {
1093 if (!is->seek_req) {
1094 is->seek_pos = pos;
1095 is->seek_rel = rel;
1096 is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1097 if (seek_by_bytes)
1098 is->seek_flags |= AVSEEK_FLAG_BYTE;
1099 is->seek_req = 1;
1100 }
1101 }
1102
1103 /* pause or resume the video */
1104 static void stream_pause(VideoState *is)
1105 {
1106 if (is->paused) {
1107 is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1108 if(is->read_pause_return != AVERROR(ENOSYS)){
1109 is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1110 }
1111 is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1112 }
1113 is->paused = !is->paused;
1114 }
1115
1116 static double compute_target_time(double frame_current_pts, VideoState *is)
1117 {
1118 double delay, sync_threshold, diff;
1119
1120 /* compute nominal delay */
1121 delay = frame_current_pts - is->frame_last_pts;
1122 if (delay <= 0 || delay >= 10.0) {
1123 /* if incorrect delay, use previous one */
1124 delay = is->frame_last_delay;
1125 } else {
1126 is->frame_last_delay = delay;
1127 }
1128 is->frame_last_pts = frame_current_pts;
1129
1130 /* update delay to follow master synchronisation source */
1131 if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1132 is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1133 /* if video is slave, we try to correct big delays by
1134 duplicating or deleting a frame */
1135 diff = get_video_clock(is) - get_master_clock(is);
1136
1137 /* skip or repeat frame. We take into account the
1138 delay to compute the threshold. I still don't know
1139 if it is the best guess */
1140 sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1141 if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1142 if (diff <= -sync_threshold)
1143 delay = 0;
1144 else if (diff >= sync_threshold)
1145 delay = 2 * delay;
1146 }
1147 }
1148 is->frame_timer += delay;
1149 #if defined(DEBUG_SYNC)
1150 printf("video: delay=%0.3f actual_delay=%0.3f pts=%0.3f A-V=%f\n",
1151 delay, actual_delay, frame_current_pts, -diff);
1152 #endif
1153
1154 return is->frame_timer;
1155 }
1156
1157 /* called to display each frame */
1158 static void video_refresh_timer(void *opaque)
1159 {
1160 VideoState *is = opaque;
1161 VideoPicture *vp;
1162
1163 SubPicture *sp, *sp2;
1164
1165 if (is->video_st) {
1166 retry:
1167 if (is->pictq_size == 0) {
1168 //nothing to do, no picture to display in the que
1169 } else {
1170 double time= av_gettime()/1000000.0;
1171 double next_target;
1172 /* dequeue the picture */
1173 vp = &is->pictq[is->pictq_rindex];
1174
1175 if(time < vp->target_clock)
1176 return;
1177 /* update current video pts */
1178 is->video_current_pts = vp->pts;
1179 is->video_current_pts_drift = is->video_current_pts - time;
1180 is->video_current_pos = vp->pos;
1181 if(is->pictq_size > 1){
1182 VideoPicture *nextvp= &is->pictq[(is->pictq_rindex+1)%VIDEO_PICTURE_QUEUE_SIZE];
1183 assert(nextvp->target_clock >= vp->target_clock);
1184 next_target= nextvp->target_clock;
1185 }else{
1186 next_target= vp->target_clock + is->video_clock - vp->pts; //FIXME pass durations cleanly
1187 }
1188 if(framedrop && time > next_target){
1189 is->skip_frames *= 1.0 + FRAME_SKIP_FACTOR;
1190 if(is->pictq_size > 1 || time > next_target + 0.5){
1191 /* update queue size and signal for next picture */
1192 if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1193 is->pictq_rindex = 0;
1194
1195 SDL_LockMutex(is->pictq_mutex);
1196 is->pictq_size--;
1197 SDL_CondSignal(is->pictq_cond);
1198 SDL_UnlockMutex(is->pictq_mutex);
1199 goto retry;
1200 }
1201 }
1202
1203 if(is->subtitle_st) {
1204 if (is->subtitle_stream_changed) {
1205 SDL_LockMutex(is->subpq_mutex);
1206
1207 while (is->subpq_size) {
1208 free_subpicture(&is->subpq[is->subpq_rindex]);
1209
1210 /* update queue size and signal for next picture */
1211 if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1212 is->subpq_rindex = 0;
1213
1214 is->subpq_size--;
1215 }
1216 is->subtitle_stream_changed = 0;
1217
1218 SDL_CondSignal(is->subpq_cond);
1219 SDL_UnlockMutex(is->subpq_mutex);
1220 } else {
1221 if (is->subpq_size > 0) {
1222 sp = &is->subpq[is->subpq_rindex];
1223
1224 if (is->subpq_size > 1)
1225 sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1226 else
1227 sp2 = NULL;
1228
1229 if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1230 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1231 {
1232 free_subpicture(sp);
1233
1234 /* update queue size and signal for next picture */
1235 if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1236 is->subpq_rindex = 0;
1237
1238 SDL_LockMutex(is->subpq_mutex);
1239 is->subpq_size--;
1240 SDL_CondSignal(is->subpq_cond);
1241 SDL_UnlockMutex(is->subpq_mutex);
1242 }
1243 }
1244 }
1245 }
1246
1247 /* display picture */
1248 video_display(is);
1249
1250 /* update queue size and signal for next picture */
1251 if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1252 is->pictq_rindex = 0;
1253
1254 SDL_LockMutex(is->pictq_mutex);
1255 is->pictq_size--;
1256 SDL_CondSignal(is->pictq_cond);
1257 SDL_UnlockMutex(is->pictq_mutex);
1258 }
1259 } else if (is->audio_st) {
1260 /* draw the next audio frame */
1261
1262 /* if only audio stream, then display the audio bars (better
1263 than nothing, just to test the implementation */
1264
1265 /* display picture */
1266 video_display(is);
1267 }
1268 if (show_status) {
1269 static int64_t last_time;
1270 int64_t cur_time;
1271 int aqsize, vqsize, sqsize;
1272 double av_diff;
1273
1274 cur_time = av_gettime();
1275 if (!last_time || (cur_time - last_time) >= 30000) {
1276 aqsize = 0;
1277 vqsize = 0;
1278 sqsize = 0;
1279 if (is->audio_st)
1280 aqsize = is->audioq.size;
1281 if (is->video_st)
1282 vqsize = is->videoq.size;
1283 if (is->subtitle_st)
1284 sqsize = is->subtitleq.size;
1285 av_diff = 0;
1286 if (is->audio_st && is->video_st)
1287 av_diff = get_audio_clock(is) - get_video_clock(is);
1288 printf("%7.2f A-V:%7.3f s:%3.1f aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64" \r",
1289 get_master_clock(is), av_diff, FFMAX(is->skip_frames-1, 0), aqsize / 1024, vqsize / 1024, sqsize, is->faulty_dts, is->faulty_pts);
1290 fflush(stdout);
1291 last_time = cur_time;
1292 }
1293 }
1294 }
1295
1296 /* allocate a picture (needs to do that in main thread to avoid
1297 potential locking problems */
1298 static void alloc_picture(void *opaque)
1299 {
1300 VideoState *is = opaque;
1301 VideoPicture *vp;
1302
1303 vp = &is->pictq[is->pictq_windex];
1304
1305 if (vp->bmp)
1306 SDL_FreeYUVOverlay(vp->bmp);
1307
1308 #if CONFIG_AVFILTER
1309 if (vp->picref)
1310 avfilter_unref_pic(vp->picref);
1311 vp->picref = NULL;
1312
1313 vp->width = is->out_video_filter->inputs[0]->w;
1314 vp->height = is->out_video_filter->inputs[0]->h;
1315 vp->pix_fmt = is->out_video_filter->inputs[0]->format;
1316 #else
1317 vp->width = is->video_st->codec->width;
1318 vp->height = is->video_st->codec->height;
1319 vp->pix_fmt = is->video_st->codec->pix_fmt;
1320 #endif
1321
1322 vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1323 SDL_YV12_OVERLAY,
1324 screen);
1325
1326 SDL_LockMutex(is->pictq_mutex);
1327 vp->allocated = 1;
1328 SDL_CondSignal(is->pictq_cond);
1329 SDL_UnlockMutex(is->pictq_mutex);
1330 }
1331
1332 /**
1333 *
1334 * @param pts the dts of the pkt / pts of the frame and guessed if not known
1335 */
1336 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts, int64_t pos)
1337 {
1338 VideoPicture *vp;
1339 int dst_pix_fmt;
1340 #if CONFIG_AVFILTER
1341 AVPicture pict_src;
1342 #endif
1343 /* wait until we have space to put a new picture */
1344 SDL_LockMutex(is->pictq_mutex);
1345
1346 if(is->pictq_size>=VIDEO_PICTURE_QUEUE_SIZE && !is->refresh)
1347 is->skip_frames= FFMAX(1.0 - FRAME_SKIP_FACTOR, is->skip_frames * (1.0-FRAME_SKIP_FACTOR));
1348
1349 while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1350 !is->videoq.abort_request) {
1351 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1352 }
1353 SDL_UnlockMutex(is->pictq_mutex);
1354
1355 if (is->videoq.abort_request)
1356 return -1;
1357
1358 vp = &is->pictq[is->pictq_windex];
1359
1360 /* alloc or resize hardware picture buffer */
1361 if (!vp->bmp ||
1362 #if CONFIG_AVFILTER
1363 vp->width != is->out_video_filter->inputs[0]->w ||
1364 vp->height != is->out_video_filter->inputs[0]->h) {
1365 #else
1366 vp->width != is->video_st->codec->width ||
1367 vp->height != is->video_st->codec->height) {
1368 #endif
1369 SDL_Event event;
1370
1371 vp->allocated = 0;
1372
1373 /* the allocation must be done in the main thread to avoid
1374 locking problems */
1375 event.type = FF_ALLOC_EVENT;
1376 event.user.data1 = is;
1377 SDL_PushEvent(&event);
1378
1379 /* wait until the picture is allocated */
1380 SDL_LockMutex(is->pictq_mutex);
1381 while (!vp->allocated && !is->videoq.abort_request) {
1382 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1383 }
1384 SDL_UnlockMutex(is->pictq_mutex);
1385
1386 if (is->videoq.abort_request)
1387 return -1;
1388 }
1389
1390 /* if the frame is not skipped, then display it */
1391 if (vp->bmp) {
1392 AVPicture pict;
1393 #if CONFIG_AVFILTER
1394 if(vp->picref)
1395 avfilter_unref_pic(vp->picref);
1396 vp->picref = src_frame->opaque;
1397 #endif
1398
1399 /* get a pointer on the bitmap */
1400 SDL_LockYUVOverlay (vp->bmp);
1401
1402 dst_pix_fmt = PIX_FMT_YUV420P;
1403 memset(&pict,0,sizeof(AVPicture));
1404 pict.data[0] = vp->bmp->pixels[0];
1405 pict.data[1] = vp->bmp->pixels[2];
1406 pict.data[2] = vp->bmp->pixels[1];
1407
1408 pict.linesize[0] = vp->bmp->pitches[0];
1409 pict.linesize[1] = vp->bmp->pitches[2];
1410 pict.linesize[2] = vp->bmp->pitches[1];
1411
1412 #if CONFIG_AVFILTER
1413 pict_src.data[0] = src_frame->data[0];
1414 pict_src.data[1] = src_frame->data[1];
1415 pict_src.data[2] = src_frame->data[2];
1416
1417 pict_src.linesize[0] = src_frame->linesize[0];
1418 pict_src.linesize[1] = src_frame->linesize[1];
1419 pict_src.linesize[2] = src_frame->linesize[2];
1420
1421 //FIXME use direct rendering
1422 av_picture_copy(&pict, &pict_src,
1423 vp->pix_fmt, vp->width, vp->height);
1424 #else
1425 sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1426 is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1427 vp->width, vp->height, vp->pix_fmt, vp->width, vp->height,
1428 dst_pix_fmt, sws_flags, NULL, NULL, NULL);
1429 if (is->img_convert_ctx == NULL) {
1430 fprintf(stderr, "Cannot initialize the conversion context\n");
1431 exit(1);
1432 }
1433 sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1434 0, vp->height, pict.data, pict.linesize);
1435 #endif
1436 /* update the bitmap content */
1437 SDL_UnlockYUVOverlay(vp->bmp);
1438
1439 vp->pts = pts;
1440 vp->pos = pos;
1441
1442 /* now we can update the picture count */
1443 if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1444 is->pictq_windex = 0;
1445 SDL_LockMutex(is->pictq_mutex);
1446 vp->target_clock= compute_target_time(vp->pts, is);
1447
1448 is->pictq_size++;
1449 SDL_UnlockMutex(is->pictq_mutex);
1450 }
1451 return 0;
1452 }
1453
1454 /**
1455 * compute the exact PTS for the picture if it is omitted in the stream
1456 * @param pts1 the dts of the pkt / pts of the frame
1457 */
1458 static int output_picture2(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1459 {
1460 double frame_delay, pts;
1461
1462 pts = pts1;
1463
1464 if (pts != 0) {
1465 /* update video clock with pts, if present */
1466 is->video_clock = pts;
1467 } else {
1468 pts = is->video_clock;
1469 }
1470 /* update video clock for next frame */
1471 frame_delay = av_q2d(is->video_st->codec->time_base);
1472 /* for MPEG2, the frame can be repeated, so we update the
1473 clock accordingly */
1474 frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1475 is->video_clock += frame_delay;
1476
1477 #if defined(DEBUG_SYNC) && 0
1478 printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1479 av_get_pict_type_char(src_frame->pict_type), pts, pts1);
1480 #endif
1481 return queue_picture(is, src_frame, pts, pos);
1482 }
1483
1484 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1485 {
1486 int len1, got_picture, i;
1487
1488 if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1489 return -1;
1490
1491 if(pkt->data == flush_pkt.data){
1492 avcodec_flush_buffers(is->video_st->codec);
1493
1494 SDL_LockMutex(is->pictq_mutex);
1495 //Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1496 for(i=0; i<VIDEO_PICTURE_QUEUE_SIZE; i++){
1497 is->pictq[i].target_clock= 0;
1498 }
1499 while (is->pictq_size && !is->videoq.abort_request) {
1500 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1501 }
1502 is->video_current_pos= -1;
1503 SDL_UnlockMutex(is->pictq_mutex);
1504
1505 is->last_dts_for_fault_detection=
1506 is->last_pts_for_fault_detection= INT64_MIN;
1507 is->frame_last_pts= AV_NOPTS_VALUE;
1508 is->frame_last_delay = 0;
1509 is->frame_timer = (double)av_gettime() / 1000000.0;
1510 is->skip_frames= 1;
1511 is->skip_frames_index= 0;
1512 return 0;
1513 }
1514
1515 /* NOTE: ipts is the PTS of the _first_ picture beginning in
1516 this packet, if any */
1517 is->video_st->codec->reordered_opaque= pkt->pts;
1518 len1 = avcodec_decode_video2(is->video_st->codec,
1519 frame, &got_picture,
1520 pkt);
1521
1522 if (got_picture) {
1523 if(pkt->dts != AV_NOPTS_VALUE){
1524 is->faulty_dts += pkt->dts <= is->last_dts_for_fault_detection;
1525 is->last_dts_for_fault_detection= pkt->dts;
1526 }
1527 if(frame->reordered_opaque != AV_NOPTS_VALUE){
1528 is->faulty_pts += frame->reordered_opaque <= is->last_pts_for_fault_detection;
1529 is->last_pts_for_fault_detection= frame->reordered_opaque;
1530 }
1531 }
1532
1533 if( ( decoder_reorder_pts==1
1534 || (decoder_reorder_pts && is->faulty_pts<is->faulty_dts)
1535 || pkt->dts == AV_NOPTS_VALUE)
1536 && frame->reordered_opaque != AV_NOPTS_VALUE)
1537 *pts= frame->reordered_opaque;
1538 else if(pkt->dts != AV_NOPTS_VALUE)
1539 *pts= pkt->dts;
1540 else
1541 *pts= 0;
1542
1543 // if (len1 < 0)
1544 // break;
1545 if (got_picture){
1546 is->skip_frames_index += 1;
1547 if(is->skip_frames_index >= is->skip_frames){
1548 is->skip_frames_index -= FFMAX(is->skip_frames, 1.0);
1549 return 1;
1550 }
1551
1552 }
1553 return 0;
1554 }
1555
1556 #if CONFIG_AVFILTER
1557 typedef struct {
1558 VideoState *is;
1559 AVFrame *frame;
1560 int use_dr1;
1561 } FilterPriv;
1562
1563 static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
1564 {
1565 AVFilterContext *ctx = codec->opaque;
1566 AVFilterPicRef *ref;
1567 int perms = AV_PERM_WRITE;
1568 int w, h, stride[4];
1569 unsigned edge;
1570
1571 if(pic->buffer_hints & FF_BUFFER_HINTS_VALID) {
1572 if(pic->buffer_hints & FF_BUFFER_HINTS_READABLE) perms |= AV_PERM_READ;
1573 if(pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) perms |= AV_PERM_PRESERVE;
1574 if(pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) perms |= AV_PERM_REUSE2;
1575 }
1576 if(pic->reference) perms |= AV_PERM_READ | AV_PERM_PRESERVE;
1577
1578 w = codec->width;
1579 h = codec->height;
1580 avcodec_align_dimensions2(codec, &w, &h, stride);
1581 edge = codec->flags & CODEC_FLAG_EMU_EDGE ? 0 : avcodec_get_edge_width();
1582 w += edge << 1;
1583 h += edge << 1;
1584
1585 if(!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
1586 return -1;
1587
1588 ref->w = codec->width;
1589 ref->h = codec->height;
1590 for(int i = 0; i < 3; i ++) {
1591 unsigned hshift = i == 0 ? 0 : av_pix_fmt_descriptors[ref->pic->format].log2_chroma_w;
1592 unsigned vshift = i == 0 ? 0 : av_pix_fmt_descriptors[ref->pic->format].log2_chroma_h;
1593
1594 ref->data[i] += (edge >> hshift) + ((edge * ref->linesize[i]) >> vshift);
1595 pic->data[i] = ref->data[i];
1596 pic->linesize[i] = ref->linesize[i];
1597 }
1598 pic->opaque = ref;
1599 pic->age = INT_MAX;
1600 pic->type = FF_BUFFER_TYPE_USER;
1601 return 0;
1602 }
1603
1604 static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
1605 {
1606 memset(pic->data, 0, sizeof(pic->data));
1607 avfilter_unref_pic(pic->opaque);
1608 }
1609
1610 static int input_init(AVFilterContext *ctx, const char *args, void *opaque)
1611 {
1612 FilterPriv *priv = ctx->priv;
1613 AVCodecContext *codec;
1614 if(!opaque) return -1;
1615
1616 priv->is = opaque;
1617 codec = priv->is->video_st->codec;
1618 codec->opaque = ctx;
1619 if(codec->codec->capabilities & CODEC_CAP_DR1) {
1620 priv->use_dr1 = 1;
1621 codec->get_buffer = input_get_buffer;
1622 codec->release_buffer = input_release_buffer;
1623 }
1624
1625 priv->frame = avcodec_alloc_frame();
1626
1627 return 0;
1628 }
1629
1630 static void input_uninit(AVFilterContext *ctx)
1631 {
1632 FilterPriv *priv = ctx->priv;
1633 av_free(priv->frame);
1634 }
1635
1636 static int input_request_frame(AVFilterLink *link)
1637 {
1638 FilterPriv *priv = link->src->priv;
1639 AVFilterPicRef *picref;
1640 int64_t pts = 0;
1641 AVPacket pkt;
1642 int ret;
1643
1644 while (!(ret = get_video_frame(priv->is, priv->frame, &pts, &pkt)))
1645 av_free_packet(&pkt);
1646 if (ret < 0)
1647 return -1;
1648
1649 if(priv->use_dr1) {
1650 picref = avfilter_ref_pic(priv->frame->opaque, ~0);
1651 } else {
1652 picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
1653 av_picture_copy((AVPicture *)&picref->data, (AVPicture *)priv->frame,
1654 picref->pic->format, link->w, link->h);
1655 }
1656 av_free_packet(&pkt);
1657
1658 picref->pts = pts;
1659 picref->pos = pkt.pos;
1660 picref->pixel_aspect = priv->is->video_st->codec->sample_aspect_ratio;
1661 avfilter_start_frame(link, picref);
1662 avfilter_draw_slice(link, 0, link->h, 1);
1663 avfilter_end_frame(link);
1664
1665 return 0;
1666 }
1667
1668 static int input_query_formats(AVFilterContext *ctx)
1669 {
1670 FilterPriv *priv = ctx->priv;
1671 enum PixelFormat pix_fmts[] = {
1672 priv->is->video_st->codec->pix_fmt, PIX_FMT_NONE
1673 };
1674
1675 avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
1676 return 0;
1677 }
1678
1679 static int input_config_props(AVFilterLink *link)
1680 {
1681 FilterPriv *priv = link->src->priv;
1682 AVCodecContext *c = priv->is->video_st->codec;
1683
1684 link->w = c->width;
1685 link->h = c->height;
1686
1687 return 0;
1688 }
1689
1690 static AVFilter input_filter =
1691 {
1692 .name = "ffplay_input",
1693
1694 .priv_size = sizeof(FilterPriv),
1695
1696 .init = input_init,
1697 .uninit = input_uninit,
1698
1699 .query_formats = input_query_formats,
1700
1701 .inputs = (AVFilterPad[]) {{ .name = NULL }},
1702 .outputs = (AVFilterPad[]) {{ .name = "default",
1703 .type = AVMEDIA_TYPE_VIDEO,
1704 .request_frame = input_request_frame,
1705 .config_props = input_config_props, },
1706 { .name = NULL }},
1707 };
1708
1709 static void output_end_frame(AVFilterLink *link)
1710 {
1711 }
1712
1713 static int output_query_formats(AVFilterContext *ctx)
1714 {
1715 enum PixelFormat pix_fmts[] = { PIX_FMT_YUV420P, PIX_FMT_NONE };
1716
1717 avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
1718 return 0;
1719 }
1720
1721 static int get_filtered_video_frame(AVFilterContext *ctx, AVFrame *frame,
1722 int64_t *pts, int64_t *pos)
1723 {
1724 AVFilterPicRef *pic;
1725
1726 if(avfilter_request_frame(ctx->inputs[0]))
1727 return -1;
1728 if(!(pic = ctx->inputs[0]->cur_pic))
1729 return -1;
1730 ctx->inputs[0]->cur_pic = NULL;
1731
1732 frame->opaque = pic;
1733 *pts = pic->pts;
1734 *pos = pic->pos;
1735
1736 memcpy(frame->data, pic->data, sizeof(frame->data));
1737 memcpy(frame->linesize, pic->linesize, sizeof(frame->linesize));
1738
1739 return 1;
1740 }
1741
1742 static AVFilter output_filter =
1743 {
1744 .name = "ffplay_output",
1745
1746 .query_formats = output_query_formats,
1747
1748 .inputs = (AVFilterPad[]) {{ .name = "default",
1749 .type = AVMEDIA_TYPE_VIDEO,
1750 .end_frame = output_end_frame,
1751 .min_perms = AV_PERM_READ, },
1752 { .name = NULL }},
1753 .outputs = (AVFilterPad[]) {{ .name = NULL }},
1754 };
1755 #endif /* CONFIG_AVFILTER */
1756
1757 static int video_thread(void *arg)
1758 {
1759 VideoState *is = arg;
1760 AVFrame *frame= avcodec_alloc_frame();
1761 int64_t pts_int;
1762 double pts;
1763 int ret;
1764
1765 #if CONFIG_AVFILTER
1766 int64_t pos;
1767 char sws_flags_str[128];
1768 AVFilterContext *filt_src = NULL, *filt_out = NULL;
1769 AVFilterGraph *graph = av_mallocz(sizeof(AVFilterGraph));
1770 snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1771 graph->scale_sws_opts = av_strdup(sws_flags_str);
1772
1773 if(!(filt_src = avfilter_open(&input_filter, "src"))) goto the_end;
1774 if(!(filt_out = avfilter_open(&output_filter, "out"))) goto the_end;
1775
1776 if(avfilter_init_filter(filt_src, NULL, is)) goto the_end;
1777 if(avfilter_init_filter(filt_out, NULL, frame)) goto the_end;
1778
1779
1780 if(vfilters) {
1781 AVFilterInOut *outputs = av_malloc(sizeof(AVFilterInOut));
1782 AVFilterInOut *inputs = av_malloc(sizeof(AVFilterInOut));
1783
1784 outputs->name = av_strdup("in");
1785 outputs->filter = filt_src;
1786 outputs->pad_idx = 0;
1787 outputs->next = NULL;
1788
1789 inputs->name = av_strdup("out");
1790 inputs->filter = filt_out;
1791 inputs->pad_idx = 0;
1792 inputs->next = NULL;
1793
1794 if (avfilter_graph_parse(graph, vfilters, inputs, outputs, NULL) < 0)
1795 goto the_end;
1796 av_freep(&vfilters);
1797 } else {
1798 if(avfilter_link(filt_src, 0, filt_out, 0) < 0) goto the_end;
1799 }
1800 avfilter_graph_add_filter(graph, filt_src);
1801 avfilter_graph_add_filter(graph, filt_out);
1802
1803 if(avfilter_graph_check_validity(graph, NULL)) goto the_end;
1804 if(avfilter_graph_config_formats(graph, NULL)) goto the_end;
1805 if(avfilter_graph_config_links(graph, NULL)) goto the_end;
1806
1807 is->out_video_filter = filt_out;
1808 #endif
1809
1810 for(;;) {
1811 #if !CONFIG_AVFILTER
1812 AVPacket pkt;
1813 #endif
1814 while (is->paused && !is->videoq.abort_request)
1815 SDL_Delay(10);
1816 #if CONFIG_AVFILTER
1817 ret = get_filtered_video_frame(filt_out, frame, &pts_int, &pos);
1818 #else
1819 ret = get_video_frame(is, frame, &pts_int, &pkt);
1820 #endif
1821
1822 if (ret < 0) goto the_end;
1823
1824 if (!ret)
1825 continue;
1826
1827 pts = pts_int*av_q2d(is->video_st->time_base);
1828
1829 #if CONFIG_AVFILTER
1830 ret = output_picture2(is, frame, pts, pos);
1831 #else
1832 ret = output_picture2(is, frame, pts, pkt.pos);
1833 av_free_packet(&pkt);
1834 #endif
1835 if (ret < 0)
1836 goto the_end;
1837
1838 if (step)
1839 if (cur_stream)
1840 stream_pause(cur_stream);
1841 }
1842 the_end:
1843 #if CONFIG_AVFILTER
1844 avfilter_graph_destroy(graph);
1845 av_freep(&graph);
1846 #endif
1847 av_free(frame);
1848 return 0;
1849 }
1850
1851 static int subtitle_thread(void *arg)
1852 {
1853 VideoState *is = arg;
1854 SubPicture *sp;
1855 AVPacket pkt1, *pkt = &pkt1;
1856 int len1, got_subtitle;
1857 double pts;
1858 int i, j;
1859 int r, g, b, y, u, v, a;
1860
1861 for(;;) {
1862 while (is->paused && !is->subtitleq.abort_request) {
1863 SDL_Delay(10);
1864 }
1865 if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1866 break;
1867
1868 if(pkt->data == flush_pkt.data){
1869 avcodec_flush_buffers(is->subtitle_st->codec);
1870 continue;
1871 }
1872 SDL_LockMutex(is->subpq_mutex);
1873 while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1874 !is->subtitleq.abort_request) {
1875 SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1876 }
1877 SDL_UnlockMutex(is->subpq_mutex);
1878
1879 if (is->subtitleq.abort_request)
1880 goto the_end;
1881
1882 sp = &is->subpq[is->subpq_windex];
1883
1884 /* NOTE: ipts is the PTS of the _first_ picture beginning in
1885 this packet, if any */
1886 pts = 0;
1887 if (pkt->pts != AV_NOPTS_VALUE)
1888 pts = av_q2d(is->subtitle_st->time_base)*pkt->pts;
1889
1890 len1 = avcodec_decode_subtitle2(is->subtitle_st->codec,
1891 &sp->sub, &got_subtitle,
1892 pkt);
1893 // if (len1 < 0)
1894 // break;
1895 if (got_subtitle && sp->sub.format == 0) {
1896 sp->pts = pts;
1897
1898 for (i = 0; i < sp->sub.num_rects; i++)
1899 {
1900 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1901 {
1902 RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1903 y = RGB_TO_Y_CCIR(r, g, b);
1904 u = RGB_TO_U_CCIR(r, g, b, 0);
1905 v = RGB_TO_V_CCIR(r, g, b, 0);
1906 YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1907 }
1908 }
1909
1910 /* now we can update the picture count */
1911 if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1912 is->subpq_windex = 0;
1913 SDL_LockMutex(is->subpq_mutex);
1914 is->subpq_size++;
1915 SDL_UnlockMutex(is->subpq_mutex);
1916 }
1917 av_free_packet(pkt);
1918 // if (step)
1919 // if (cur_stream)
1920 // stream_pause(cur_stream);
1921 }
1922 the_end:
1923 return 0;
1924 }
1925
1926 /* copy samples for viewing in editor window */
1927 static void update_sample_display(VideoState *is, short *samples, int samples_size)
1928 {
1929 int size, len, channels;
1930
1931 channels = is->audio_st->codec->channels;
1932
1933 size = samples_size / sizeof(short);
1934 while (size > 0) {
1935 len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
1936 if (len > size)
1937 len = size;
1938 memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
1939 samples += len;
1940 is->sample_array_index += len;
1941 if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
1942 is->sample_array_index = 0;
1943 size -= len;
1944 }
1945 }
1946
1947 /* return the new audio buffer size (samples can be added or deleted
1948 to get better sync if video or external master clock) */
1949 static int synchronize_audio(VideoState *is, short *samples,
1950 int samples_size1, double pts)
1951 {
1952 int n, samples_size;
1953 double ref_clock;
1954
1955 n = 2 * is->audio_st->codec->channels;
1956 samples_size = samples_size1;
1957
1958 /* if not master, then we try to remove or add samples to correct the clock */
1959 if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
1960 is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1961 double diff, avg_diff;
1962 int wanted_size, min_size, max_size, nb_samples;
1963
1964 ref_clock = get_master_clock(is);
1965 diff = get_audio_clock(is) - ref_clock;
1966
1967 if (diff < AV_NOSYNC_THRESHOLD) {
1968 is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
1969 if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
1970 /* not enough measures to have a correct estimate */
1971 is->audio_diff_avg_count++;
1972 } else {
1973 /* estimate the A-V difference */
1974 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
1975
1976 if (fabs(avg_diff) >= is->audio_diff_threshold) {
1977 wanted_size = samples_size + ((int)(diff * is->audio_st->codec->sample_rate) * n);
1978 nb_samples = samples_size / n;
1979
1980 min_size = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
1981 max_size = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
1982 if (wanted_size < min_size)
1983 wanted_size = min_size;
1984 else if (wanted_size > max_size)
1985 wanted_size = max_size;
1986
1987 /* add or remove samples to correction the synchro */
1988 if (wanted_size < samples_size) {
1989 /* remove samples */
1990 samples_size = wanted_size;
1991 } else if (wanted_size > samples_size) {
1992 uint8_t *samples_end, *q;
1993 int nb;
1994
1995 /* add samples */
1996 nb = (samples_size - wanted_size);
1997 samples_end = (uint8_t *)samples + samples_size - n;
1998 q = samples_end + n;
1999 while (nb > 0) {
2000 memcpy(q, samples_end, n);
2001 q += n;
2002 nb -= n;
2003 }
2004 samples_size = wanted_size;
2005 }
2006 }
2007 #if 0
2008 printf("diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
2009 diff, avg_diff, samples_size - samples_size1,
2010 is->audio_clock, is->video_clock, is->audio_diff_threshold);
2011 #endif
2012 }
2013 } else {
2014 /* too big difference : may be initial PTS errors, so
2015 reset A-V filter */
2016 is->audio_diff_avg_count = 0;
2017 is->audio_diff_cum = 0;
2018 }
2019 }
2020
2021 return samples_size;
2022 }
2023
2024 /* decode one audio frame and returns its uncompressed size */
2025 static int audio_decode_frame(VideoState *is, double *pts_ptr)
2026 {
2027 AVPacket *pkt_temp = &is->audio_pkt_temp;
2028 AVPacket *pkt = &is->audio_pkt;
2029 AVCodecContext *dec= is->audio_st->codec;
2030 int n, len1, data_size;
2031 double pts;
2032
2033 for(;;) {
2034 /* NOTE: the audio packet can contain several frames */
2035 while (pkt_temp->size > 0) {
2036 data_size = sizeof(is->audio_buf1);
2037 len1 = avcodec_decode_audio3(dec,
2038 (int16_t *)is->audio_buf1, &data_size,
2039 pkt_temp);
2040 if (len1 < 0) {
2041 /* if error, we skip the frame */
2042 pkt_temp->size = 0;
2043 break;
2044 }
2045
2046 pkt_temp->data += len1;
2047 pkt_temp->size -= len1;
2048 if (data_size <= 0)
2049 continue;
2050
2051 if (dec->sample_fmt != is->audio_src_fmt) {
2052 if (is->reformat_ctx)
2053 av_audio_convert_free(is->reformat_ctx);
2054 is->reformat_ctx= av_audio_convert_alloc(SAMPLE_FMT_S16, 1,
2055 dec->sample_fmt, 1, NULL, 0);
2056 if (!is->reformat_ctx) {
2057 fprintf(stderr, "Cannot convert %s sample format to %s sample format\n",
2058 avcodec_get_sample_fmt_name(dec->sample_fmt),
2059 avcodec_get_sample_fmt_name(SAMPLE_FMT_S16));
2060 break;
2061 }
2062 is->audio_src_fmt= dec->sample_fmt;
2063 }
2064
2065 if (is->reformat_ctx) {
2066 const void *ibuf[6]= {is->audio_buf1};
2067 void *obuf[6]= {is->audio_buf2};
2068 int istride[6]= {av_get_bits_per_sample_format(dec->sample_fmt)/8};
2069 int ostride[6]= {2};
2070 int len= data_size/istride[0];
2071 if (av_audio_convert(is->reformat_ctx, obuf, ostride, ibuf, istride, len)<0) {
2072 printf("av_audio_convert() failed\n");
2073 break;
2074 }
2075 is->audio_buf= is->audio_buf2;
2076 /* FIXME: existing code assume that data_size equals framesize*channels*2
2077 remove this legacy cruft */
2078 data_size= len*2;
2079 }else{
2080 is->audio_buf= is->audio_buf1;
2081 }
2082
2083 /* if no pts, then compute it */
2084 pts = is->audio_clock;
2085 *pts_ptr = pts;
2086 n = 2 * dec->channels;
2087 is->audio_clock += (double)data_size /
2088 (double)(n * dec->sample_rate);
2089 #if defined(DEBUG_SYNC)
2090 {
2091 static double last_clock;
2092 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2093 is->audio_clock - last_clock,
2094 is->audio_clock, pts);
2095 last_clock = is->audio_clock;
2096 }
2097 #endif
2098 return data_size;
2099 }
2100
2101 /* free the current packet */
2102 if (pkt->data)
2103 av_free_packet(pkt);
2104
2105 if (is->paused || is->audioq.abort_request) {
2106 return -1;
2107 }
2108
2109 /* read next packet */
2110 if (packet_queue_get(&is->audioq, pkt, 1) < 0)
2111 return -1;
2112 if(pkt->data == flush_pkt.data){
2113 avcodec_flush_buffers(dec);
2114 continue;
2115 }
2116
2117 pkt_temp->data = pkt->data;
2118 pkt_temp->size = pkt->size;
2119
2120 /* if update the audio clock with the pts */
2121 if (pkt->pts != AV_NOPTS_VALUE) {
2122 is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2123 }
2124 }
2125 }
2126
2127 /* get the current audio output buffer size, in samples. With SDL, we
2128 cannot have a precise information */
2129 static int audio_write_get_buf_size(VideoState *is)
2130 {
2131 return is->audio_buf_size - is->audio_buf_index;
2132 }
2133
2134
2135 /* prepare a new audio buffer */
2136 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2137 {
2138 VideoState *is = opaque;
2139 int audio_size, len1;
2140 double pts;
2141
2142 audio_callback_time = av_gettime();
2143
2144 while (len > 0) {
2145 if (is->audio_buf_index >= is->audio_buf_size) {
2146 audio_size = audio_decode_frame(is, &pts);
2147 if (audio_size < 0) {
2148 /* if error, just output silence */
2149 is->audio_buf = is->audio_buf1;
2150 is->audio_buf_size = 1024;
2151 memset(is->audio_buf, 0, is->audio_buf_size);
2152 } else {
2153 if (is->show_audio)
2154 update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2155 audio_size = synchronize_audio(is, (int16_t *)is->audio_buf, audio_size,
2156 pts);
2157 is->audio_buf_size = audio_size;
2158 }
2159 is->audio_buf_index = 0;
2160 }
2161 len1 = is->audio_buf_size - is->audio_buf_index;
2162 if (len1 > len)
2163 len1 = len;
2164 memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2165 len -= len1;
2166 stream += len1;
2167 is->audio_buf_index += len1;
2168 }
2169 }
2170
2171 /* open a given stream. Return 0 if OK */
2172 static int stream_component_open(VideoState *is, int stream_index)
2173 {
2174 AVFormatContext *ic = is->ic;
2175 AVCodecContext *avctx;
2176 AVCodec *codec;
2177 SDL_AudioSpec wanted_spec, spec;
2178
2179 if (stream_index < 0 || stream_index >= ic->nb_streams)
2180 return -1;
2181 avctx = ic->streams[stream_index]->codec;
2182
2183 /* prepare audio output */
2184 if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2185 if (avctx->channels > 0) {
2186 avctx->request_channels = FFMIN(2, avctx->channels);
2187 } else {
2188 avctx->request_channels = 2;
2189 }
2190 }
2191
2192 codec = avcodec_find_decoder(avctx->codec_id);
2193 avctx->debug_mv = debug_mv;
2194 avctx->debug = debug;
2195 avctx->workaround_bugs = workaround_bugs;
2196 avctx->lowres = lowres;
2197 if(lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2198 avctx->idct_algo= idct;
2199 if(fast) avctx->flags2 |= CODEC_FLAG2_FAST;
2200 avctx->skip_frame= skip_frame;
2201 avctx->skip_idct= skip_idct;
2202 avctx->skip_loop_filter= skip_loop_filter;
2203 avctx->error_recognition= error_recognition;
2204 avctx->error_concealment= error_concealment;
2205 avcodec_thread_init(avctx, thread_count);
2206
2207 set_context_opts(avctx, avcodec_opts[avctx->codec_type], 0);
2208
2209 if (!codec ||
2210 avcodec_open(avctx, codec) < 0)
2211 return -1;
2212
2213 /* prepare audio output */
2214 if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2215 wanted_spec.freq = avctx->sample_rate;
2216 wanted_spec.format = AUDIO_S16SYS;
2217 wanted_spec.channels = avctx->channels;
2218 wanted_spec.silence = 0;
2219 wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2220 wanted_spec.callback = sdl_audio_callback;
2221 wanted_spec.userdata = is;
2222 if (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2223 fprintf(stderr, "SDL_OpenAudio: %s\n", SDL_GetError());
2224 return -1;
2225 }
2226 is->audio_hw_buf_size = spec.size;
2227 is->audio_src_fmt= SAMPLE_FMT_S16;
2228 }
2229
2230 ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2231 switch(avctx->codec_type) {
2232 case AVMEDIA_TYPE_AUDIO:
2233 is->audio_stream = stream_index;
2234 is->audio_st = ic->streams[stream_index];
2235 is->audio_buf_size = 0;
2236 is->audio_buf_index = 0;
2237
2238 /* init averaging filter */
2239 is->audio_diff_avg_coef = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2240 is->audio_diff_avg_count = 0;
2241 /* since we do not have a precise anough audio fifo fullness,
2242 we correct audio sync only if larger than this threshold */
2243 is->audio_diff_threshold = 2.0 * SDL_AUDIO_BUFFER_SIZE / avctx->sample_rate;
2244
2245 memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2246 packet_queue_init(&is->audioq);
2247 SDL_PauseAudio(0);
2248 break;
2249 case AVMEDIA_TYPE_VIDEO:
2250 is->video_stream = stream_index;
2251 is->video_st = ic->streams[stream_index];
2252
2253 // is->video_current_pts_time = av_gettime();
2254
2255 packet_queue_init(&is->videoq);
2256 is->video_tid = SDL_CreateThread(video_thread, is);
2257 break;
2258 case AVMEDIA_TYPE_SUBTITLE:
2259 is->subtitle_stream = stream_index;
2260 is->subtitle_st = ic->streams[stream_index];
2261 packet_queue_init(&is->subtitleq);
2262
2263 is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2264 break;
2265 default:
2266 break;
2267 }
2268 return 0;
2269 }
2270
2271 static void stream_component_close(VideoState *is, int stream_index)
2272 {
2273 AVFormatContext *ic = is->ic;
2274 AVCodecContext *avctx;
2275
2276 if (stream_index < 0 || stream_index >= ic->nb_streams)
2277 return;
2278 avctx = ic->streams[stream_index]->codec;
2279
2280 switch(avctx->codec_type) {
2281 case AVMEDIA_TYPE_AUDIO:
2282 packet_queue_abort(&is->audioq);
2283
2284 SDL_CloseAudio();
2285
2286 packet_queue_end(&is->audioq);
2287 if (is->reformat_ctx)
2288 av_audio_convert_free(is->reformat_ctx);
2289 is->reformat_ctx = NULL;
2290 break;
2291 case AVMEDIA_TYPE_VIDEO:
2292 packet_queue_abort(&is->videoq);
2293
2294 /* note: we also signal this mutex to make sure we deblock the
2295 video thread in all cases */
2296 SDL_LockMutex(is->pictq_mutex);
2297 SDL_CondSignal(is->pictq_cond);
2298 SDL_UnlockMutex(is->pictq_mutex);
2299
2300 SDL_WaitThread(is->video_tid, NULL);
2301
2302 packet_queue_end(&is->videoq);
2303 break;
2304 case AVMEDIA_TYPE_SUBTITLE:
2305 packet_queue_abort(&is->subtitleq);
2306
2307 /* note: we also signal this mutex to make sure we deblock the
2308 video thread in all cases */
2309 SDL_LockMutex(is->subpq_mutex);
2310 is->subtitle_stream_changed = 1;
2311
2312 SDL_CondSignal(is->subpq_cond);
2313 SDL_UnlockMutex(is->subpq_mutex);
2314
2315 SDL_WaitThread(is->subtitle_tid, NULL);
2316
2317 packet_queue_end(&is->subtitleq);
2318 break;
2319 default:
2320 break;
2321 }
2322
2323 ic->streams[stream_index]->discard = AVDISCARD_ALL;
2324 avcodec_close(avctx);
2325 switch(avctx->codec_type) {
2326 case AVMEDIA_TYPE_AUDIO:
2327 is->audio_st = NULL;
2328 is->audio_stream = -1;
2329 break;
2330 case AVMEDIA_TYPE_VIDEO:
2331 is->video_st = NULL;
2332 is->video_stream = -1;
2333 break;
2334 case AVMEDIA_TYPE_SUBTITLE:
2335 is->subtitle_st = NULL;
2336 is->subtitle_stream = -1;
2337 break;
2338 default:
2339 break;
2340 }
2341 }
2342
2343 /* since we have only one decoding thread, we can use a global
2344 variable instead of a thread local variable */
2345 static VideoState *global_video_state;
2346
2347 static int decode_interrupt_cb(void)
2348 {
2349 return (global_video_state && global_video_state->abort_request);
2350 }
2351
2352 /* this thread gets the stream from the disk or the network */
2353 static int decode_thread(void *arg)
2354 {
2355 VideoState *is = arg;
2356 AVFormatContext *ic;
2357 int err, i, ret;
2358 int st_index[AVMEDIA_TYPE_NB];
2359 int st_count[AVMEDIA_TYPE_NB]={0};
2360 int st_best_packet_count[AVMEDIA_TYPE_NB];
2361 AVPacket pkt1, *pkt = &pkt1;
2362 AVFormatParameters params, *ap = &params;
2363 int eof=0;
2364 int pkt_in_play_range = 0;
2365
2366 ic = avformat_alloc_context();
2367
2368 memset(st_index, -1, sizeof(st_index));
2369 memset(st_best_packet_count, -1, sizeof(st_best_packet_count));
2370 is->video_stream = -1;
2371 is->audio_stream = -1;
2372 is->subtitle_stream = -1;
2373
2374 global_video_state = is;
2375 url_set_interrupt_cb(decode_interrupt_cb);
2376
2377 memset(ap, 0, sizeof(*ap));
2378
2379 ap->prealloced_context = 1;
2380 ap->width = frame_width;
2381 ap->height= frame_height;
2382 ap->time_base= (AVRational){1, 25};
2383 ap->pix_fmt = frame_pix_fmt;
2384
2385 set_context_opts(ic, avformat_opts, AV_OPT_FLAG_DECODING_PARAM);
2386
2387 err = av_open_input_file(&ic, is->filename, is->iformat, 0, ap);
2388 if (err < 0) {
2389 print_error(is->filename, err);
2390 ret = -1;
2391 goto fail;
2392 }
2393 is->ic = ic;
2394
2395 if(genpts)
2396 ic->flags |= AVFMT_FLAG_GENPTS;
2397
2398 err = av_find_stream_info(ic);
2399 if (err < 0) {
2400 fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2401 ret = -1;
2402 goto fail;
2403 }
2404 if(ic->pb)
2405 ic->pb->eof_reached= 0; //FIXME hack, ffplay maybe should not use url_feof() to test for the end
2406
2407 if(seek_by_bytes<0)
2408 seek_by_bytes= !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2409
2410 /* if seeking requested, we execute it */
2411 if (start_time != AV_NOPTS_VALUE) {
2412 int64_t timestamp;
2413
2414 timestamp = start_time;
2415 /* add the stream start time */
2416 if (ic->start_time != AV_NOPTS_VALUE)
2417 timestamp += ic->start_time;
2418 ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2419 if (ret < 0) {
2420 fprintf(stderr, "%s: could not seek to position %0.3f\n",
2421 is->filename, (double)timestamp / AV_TIME_BASE);
2422 }
2423 }
2424
2425 for(i = 0; i < ic->nb_streams; i++) {
2426 AVStream *st= ic->streams[i];
2427 AVCodecContext *avctx = st->codec;
2428 ic->streams[i]->discard = AVDISCARD_ALL;
2429 if(avctx->codec_type >= (unsigned)AVMEDIA_TYPE_NB)
2430 continue;
2431 if(st_count[avctx->codec_type]++ != wanted_stream[avctx->codec_type] && wanted_stream[avctx->codec_type] >= 0)
2432 continue;
2433
2434 if(st_best_packet_count[avctx->codec_type] >= st->codec_info_nb_frames)
2435 continue;
2436 st_best_packet_count[avctx->codec_type]= st->codec_info_nb_frames;
2437
2438 switch(avctx->codec_type) {
2439 case AVMEDIA_TYPE_AUDIO:
2440 if (!audio_disable)
2441 st_index[AVMEDIA_TYPE_AUDIO] = i;
2442 break;
2443 case AVMEDIA_TYPE_VIDEO:
2444 case AVMEDIA_TYPE_SUBTITLE:
2445 if (!video_disable)
2446 st_index[avctx->codec_type] = i;
2447 break;
2448 default:
2449 break;
2450 }
2451 }
2452 if (show_status) {
2453 dump_format(ic, 0, is->filename, 0);
2454 }
2455
2456 /* open the streams */
2457 if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2458 stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2459 }
2460
2461 ret=-1;
2462 if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2463 ret= stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2464 }
2465 is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2466 if(ret<0) {
2467 if (!display_disable)
2468 is->show_audio = 2;
2469 }
2470
2471 if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2472 stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2473 }
2474
2475 if (is->video_stream < 0 && is->audio_stream < 0) {
2476 fprintf(stderr, "%s: could not open codecs\n", is->filename);
2477 ret = -1;
2478 goto fail;
2479 }
2480
2481 for(;;) {
2482 if (is->abort_request)
2483 break;
2484 if (is->paused != is->last_paused) {
2485 is->last_paused = is->paused;
2486 if (is->paused)
2487 is->read_pause_return= av_read_pause(ic);
2488 else
2489 av_read_play(ic);
2490 }
2491 #if CONFIG_RTSP_DEMUXER
2492 if (is->paused && !strcmp(ic->iformat->name, "rtsp")) {
2493 /* wait 10 ms to avoid trying to get another packet */
2494 /* XXX: horrible */
2495 SDL_Delay(10);
2496 continue;
2497 }
2498 #endif
2499 if (is->seek_req) {
2500 int64_t seek_target= is->seek_pos;
2501 int64_t seek_min= is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2502 int64_t seek_max= is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2503 //FIXME the +-2 is due to rounding being not done in the correct direction in generation
2504 // of the seek_pos/seek_rel variables
2505
2506 ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2507 if (ret < 0) {
2508 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2509 }else{
2510 if (is->audio_stream >= 0) {
2511 packet_queue_flush(&is->audioq);
2512 packet_queue_put(&is->audioq, &flush_pkt);
2513 }
2514 if (is->subtitle_stream >= 0) {
2515 packet_queue_flush(&is->subtitleq);
2516 packet_queue_put(&is->subtitleq, &flush_pkt);
2517 }
2518 if (is->video_stream >= 0) {
2519 packet_queue_flush(&is->videoq);
2520 packet_queue_put(&is->videoq, &flush_pkt);
2521 }
2522 }
2523 is->seek_req = 0;
2524 eof= 0;
2525 }
2526
2527 /* if the queue are full, no need to read more */
2528 if ( is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2529 || ( (is->audioq .size > MIN_AUDIOQ_SIZE || is->audio_stream<0)
2530 && (is->videoq .nb_packets > MIN_FRAMES || is->video_stream<0)
2531 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream<0))) {
2532 /* wait 10 ms */
2533 SDL_Delay(10);
2534 continue;
2535 }
2536 if(url_feof(ic->pb) || eof) {
2537 if(is->video_stream >= 0){
2538 av_init_packet(pkt);
2539 pkt->data=NULL;
2540 pkt->size=0;
2541 pkt->stream_index= is->video_stream;
2542 packet_queue_put(&is->videoq, pkt);
2543 }
2544 SDL_Delay(10);
2545 if(is->audioq.size + is->videoq.size + is->subtitleq.size ==0){
2546 if(loop!=1 && (!loop || --loop)){
2547 stream_seek(cur_stream, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2548 }else if(autoexit){
2549 ret=AVERROR_EOF;
2550 goto fail;
2551 }
2552 }
2553 continue;
2554 }
2555 ret = av_read_frame(ic, pkt);
2556 if (ret < 0) {
2557 if (ret == AVERROR_EOF)
2558 eof=1;
2559 if (url_ferror(ic->pb))
2560 break;
2561 SDL_Delay(100); /* wait for user event */
2562 continue;
2563 }
2564 /* check if packet is in play range specified by user, then queue, otherwise discard */
2565 pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2566 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2567 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2568 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0)/1000000
2569 <= ((double)duration/1000000);
2570 if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2571 packet_queue_put(&is->audioq, pkt);
2572 } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2573 packet_queue_put(&is->videoq, pkt);
2574 } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2575 packet_queue_put(&is->subtitleq, pkt);
2576 } else {
2577 av_free_packet(pkt);
2578 }
2579 }
2580 /* wait until the end */
2581 while (!is->abort_request) {
2582 SDL_Delay(100);
2583 }
2584
2585 ret = 0;
2586 fail:
2587 /* disable interrupting */
2588 global_video_state = NULL;
2589
2590 /* close each stream */
2591 if (is->audio_stream >= 0)
2592 stream_component_close(is, is->audio_stream);
2593 if (is->video_stream >= 0)
2594 stream_component_close(is, is->video_stream);
2595 if (is->subtitle_stream >= 0)
2596 stream_component_close(is, is->subtitle_stream);
2597 if (is->ic) {
2598 av_close_input_file(is->ic);
2599 is->ic = NULL; /* safety */
2600 }
2601 url_set_interrupt_cb(NULL);
2602
2603 if (ret != 0) {
2604 SDL_Event event;
2605
2606 event.type = FF_QUIT_EVENT;
2607 event.user.data1 = is;
2608 SDL_PushEvent(&event);
2609 }
2610 return 0;
2611 }
2612
2613 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2614 {
2615 VideoState *is;
2616
2617 is = av_mallocz(sizeof(VideoState));
2618 if (!is)
2619 return NULL;
2620 av_strlcpy(is->filename, filename, sizeof(is->filename));
2621 is->iformat = iformat;
2622 is->ytop = 0;
2623 is->xleft = 0;
2624
2625 /* start video display */
2626 is->pictq_mutex = SDL_CreateMutex();
2627 is->pictq_cond = SDL_CreateCond();
2628
2629 is->subpq_mutex = SDL_CreateMutex();
2630 is->subpq_cond = SDL_CreateCond();
2631
2632 is->av_sync_type = av_sync_type;
2633 is->parse_tid = SDL_CreateThread(decode_thread, is);
2634 if (!is->parse_tid) {
2635 av_free(is);
2636 return NULL;
2637 }
2638 return is;
2639 }
2640
2641 static void stream_close(VideoState *is)
2642 {
2643 VideoPicture *vp;
2644 int i;
2645 /* XXX: use a special url_shutdown call to abort parse cleanly */
2646 is->abort_request = 1;
2647 SDL_WaitThread(is->parse_tid, NULL);
2648 SDL_WaitThread(is->refresh_tid, NULL);
2649
2650 /* free all pictures */
2651 for(i=0;i<VIDEO_PICTURE_QUEUE_SIZE; i++) {
2652 vp = &is->pictq[i];
2653 #if CONFIG_AVFILTER
2654 if (vp->picref) {
2655 avfilter_unref_pic(vp->picref);
2656 vp->picref = NULL;
2657 }
2658 #endif
2659 if (vp->bmp) {
2660 SDL_FreeYUVOverlay(vp->bmp);
2661 vp->bmp = NULL;
2662 }
2663 }
2664 SDL_DestroyMutex(is->pictq_mutex);
2665 SDL_DestroyCond(is->pictq_cond);
2666 SDL_DestroyMutex(is->subpq_mutex);
2667 SDL_DestroyCond(is->subpq_cond);
2668 #if !CONFIG_AVFILTER
2669 if (is->img_convert_ctx)
2670 sws_freeContext(is->img_convert_ctx);
2671 #endif
2672 av_free(is);
2673 }
2674
2675 static void stream_cycle_channel(VideoState *is, int codec_type)
2676 {
2677 AVFormatContext *ic = is->ic;
2678 int start_index, stream_index;
2679 AVStream *st;
2680
2681 if (codec_type == AVMEDIA_TYPE_VIDEO)
2682 start_index = is->video_stream;
2683 else if (codec_type == AVMEDIA_TYPE_AUDIO)
2684 start_index = is->audio_stream;
2685 else
2686 start_index = is->subtitle_stream;
2687 if (start_index < (codec_type == AVMEDIA_TYPE_SUBTITLE ? -1 : 0))
2688 return;
2689 stream_index = start_index;
2690 for(;;) {
2691 if (++stream_index >= is->ic->nb_streams)
2692 {
2693 if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2694 {
2695 stream_index = -1;
2696 goto the_end;
2697 } else
2698 stream_index = 0;
2699 }
2700 if (stream_index == start_index)
2701 return;
2702 st = ic->streams[stream_index];
2703 if (st->codec->codec_type == codec_type) {
2704 /* check that parameters are OK */
2705 switch(codec_type) {
2706 case AVMEDIA_TYPE_AUDIO:
2707 if (st->codec->sample_rate != 0 &&
2708 st->codec->channels != 0)
2709 goto the_end;
2710 break;
2711 case AVMEDIA_TYPE_VIDEO:
2712 case AVMEDIA_TYPE_SUBTITLE:
2713 goto the_end;
2714 default:
2715 break;
2716 }
2717 }
2718 }
2719 the_end:
2720 stream_component_close(is, start_index);
2721 stream_component_open(is, stream_index);
2722 }
2723
2724
2725 static void toggle_full_screen(void)
2726 {
2727 is_full_screen = !is_full_screen;
2728 if (!fs_screen_width) {
2729 /* use default SDL method */
2730 // SDL_WM_ToggleFullScreen(screen);
2731 }
2732 video_open(cur_stream);
2733 }
2734
2735 static void toggle_pause(void)
2736 {
2737 if (cur_stream)
2738 stream_pause(cur_stream);
2739 step = 0;
2740 }
2741
2742 static void step_to_next_frame(void)
2743 {
2744 if (cur_stream) {
2745 /* if the stream is paused unpause it, then step */
2746 if (cur_stream->paused)
2747 stream_pause(cur_stream);
2748 }
2749 step = 1;
2750 }
2751
2752 static void do_exit(void)
2753 {
2754 int i;
2755 if (cur_stream) {
2756 stream_close(cur_stream);
2757 cur_stream = NULL;
2758 }
2759 for (i = 0; i < AVMEDIA_TYPE_NB; i++)
2760 av_free(avcodec_opts[i]);
2761 av_free(avformat_opts);
2762 av_free(sws_opts);
2763 #if CONFIG_AVFILTER
2764 avfilter_uninit();
2765 #endif
2766 if (show_status)
2767 printf("\n");
2768 SDL_Quit();
2769 exit(0);
2770 }
2771
2772 static void toggle_audio_display(void)
2773 {
2774 if (cur_stream) {
2775 int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2776 cur_stream->show_audio = (cur_stream->show_audio + 1) % 3;
2777 fill_rectangle(screen,
2778 cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height,
2779 bgcolor);
2780 SDL_UpdateRect(screen, cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height);
2781 }
2782 }
2783
2784 /* handle an event sent by the GUI */
2785 static void event_loop(void)
2786 {
2787 SDL_Event event;
2788 double incr, pos, frac;
2789
2790 for(;;) {
2791 double x;
2792 SDL_WaitEvent(&event);
2793 switch(event.type) {
2794 case SDL_KEYDOWN:
2795 switch(event.key.keysym.sym) {
2796 case SDLK_ESCAPE:
2797 case SDLK_q:
2798 do_exit();
2799 break;
2800 case SDLK_f:
2801 toggle_full_screen();
2802 break;
2803 case SDLK_p:
2804 case SDLK_SPACE:
2805 toggle_pause();
2806 break;
2807 case SDLK_s: //S: Step to next frame
2808 step_to_next_frame();
2809 break;
2810 case SDLK_a:
2811 if (cur_stream)
2812 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2813 break;
2814 case SDLK_v:
2815 if (cur_stream)
2816 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2817 break;
2818 case SDLK_t:
2819 if (cur_stream)
2820 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2821 break;
2822 case SDLK_w:
2823 toggle_audio_display();
2824 break;
2825 case SDLK_LEFT:
2826 incr = -10.0;
2827 goto do_seek;
2828 case SDLK_RIGHT:
2829 incr = 10.0;
2830 goto do_seek;
2831 case SDLK_UP:
2832 incr = 60.0;
2833 goto do_seek;
2834 case SDLK_DOWN:
2835 incr = -60.0;
2836 do_seek:
2837 if (cur_stream) {
2838 if (seek_by_bytes) {
2839 if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos>=0){
2840 pos= cur_stream->video_current_pos;
2841 }else if(cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos>=0){
2842 pos= cur_stream->audio_pkt.pos;
2843 }else
2844 pos = url_ftell(cur_stream->ic->pb);
2845 if (cur_stream->ic->bit_rate)
2846 incr *= cur_stream->ic->bit_rate / 8.0;
2847 else
2848 incr *= 180000.0;
2849 pos += incr;
2850 stream_seek(cur_stream, pos, incr, 1);
2851 } else {
2852 pos = get_master_clock(cur_stream);
2853 pos += incr;
2854 stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2855 }
2856 }
2857 break;
2858 default:
2859 break;
2860 }
2861 break;
2862 case SDL_MOUSEBUTTONDOWN:
2863 case SDL_MOUSEMOTION:
2864 if(event.type ==SDL_MOUSEBUTTONDOWN){
2865 x= event.button.x;
2866 }else{
2867 if(event.motion.state != SDL_PRESSED)
2868 break;
2869 x= event.motion.x;
2870 }
2871 if (cur_stream) {
2872 if(seek_by_bytes || cur_stream->ic->duration<=0){
2873 uint64_t size= url_fsize(cur_stream->ic->pb);
2874 stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2875 }else{
2876 int64_t ts;
2877 int ns, hh, mm, ss;
2878 int tns, thh, tmm, tss;
2879 tns = cur_stream->ic->duration/1000000LL;
2880 thh = tns/3600;
2881 tmm = (tns%3600)/60;
2882 tss = (tns%60);
2883 frac = x/cur_stream->width;
2884 ns = frac*tns;
2885 hh = ns/3600;
2886 mm = (ns%3600)/60;
2887 ss = (ns%60);
2888 fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d) \n", frac*100,
2889 hh, mm, ss, thh, tmm, tss);
2890 ts = frac*cur_stream->ic->duration;
2891 if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2892 ts += cur_stream->ic->start_time;
2893 stream_seek(cur_stream, ts, 0, 0);
2894 }
2895 }
2896 break;
2897 case SDL_VIDEORESIZE:
2898 if (cur_stream) {
2899 screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2900 SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2901 screen_width = cur_stream->width = event.resize.w;
2902 screen_height= cur_stream->height= event.resize.h;
2903 }
2904 break;
2905 case SDL_QUIT:
2906 case FF_QUIT_EVENT:
2907 do_exit();
2908 break;
2909 case FF_ALLOC_EVENT:
2910 video_open(event.user.data1);
2911 alloc_picture(event.user.data1);
2912 break;
2913 case FF_REFRESH_EVENT:
2914 video_refresh_timer(event.user.data1);
2915 cur_stream->refresh=0;
2916 break;
2917 default:
2918 break;
2919 }
2920 }
2921 }
2922
2923 static void opt_frame_size(const char *arg)
2924 {
2925 if (av_parse_video_frame_size(&frame_width, &frame_height, arg) < 0) {
2926 fprintf(stderr, "Incorrect frame size\n");
2927 exit(1);
2928 }
2929 if ((frame_width % 2) != 0 || (frame_height % 2) != 0) {
2930 fprintf(stderr, "Frame size must be a multiple of 2\n");
2931 exit(1);
2932 }
2933 }
2934
2935 static int opt_width(const char *opt, const char *arg)
2936 {
2937 screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2938 return 0;
2939 }
2940
2941 static int opt_height(const char *opt, const char *arg)
2942 {
2943 screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2944 return 0;
2945 }
2946
2947 static void opt_format(const char *arg)
2948 {
2949 file_iformat = av_find_input_format(arg);
2950 if (!file_iformat) {
2951 fprintf(stderr, "Unknown input format: %s\n", arg);
2952 exit(1);
2953 }
2954 }
2955
2956 static void opt_frame_pix_fmt(const char *arg)
2957 {
2958 frame_pix_fmt = av_get_pix_fmt(arg);
2959 }
2960
2961 static int opt_sync(const char *opt, const char *arg)
2962 {
2963 if (!strcmp(arg, "audio"))
2964 av_sync_type = AV_SYNC_AUDIO_MASTER;
2965 else if (!strcmp(arg, "video"))
2966 av_sync_type = AV_SYNC_VIDEO_MASTER;
2967 else if (!strcmp(arg, "ext"))
2968 av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
2969 else {
2970 fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
2971 exit(1);
2972 }
2973 return 0;
2974 }
2975
2976 static int opt_seek(const char *opt, const char *arg)
2977 {
2978 start_time = parse_time_or_die(opt, arg, 1);
2979 return 0;
2980 }
2981
2982 static int opt_duration(const char *opt, const char *arg)
2983 {
2984 duration = parse_time_or_die(opt, arg, 1);
2985 return 0;
2986 }
2987
2988 static int opt_debug(const char *opt, const char *arg)
2989 {
2990 av_log_set_level(99);
2991 debug = parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
2992 return 0;
2993 }
2994
2995 static int opt_vismv(const char *opt, const char *arg)
2996 {
2997 debug_mv = parse_number_or_die(opt, arg, OPT_INT64, INT_MIN, INT_MAX);
2998 return 0;
2999 }
3000
3001 static int opt_thread_count(const char *opt, const char *arg)
3002 {
3003 thread_count= parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
3004 #if !HAVE_THREADS
3005 fprintf(stderr, "Warning: not compiled with thread support, using thread emulation\n");
3006 #endif
3007 return 0;
3008 }
3009
3010 static const OptionDef options[] = {
3011 #include "cmdutils_common_opts.h"
3012 { "x", HAS_ARG | OPT_FUNC2, {(void*)opt_width}, "force displayed width", "width" },
3013 { "y", HAS_ARG | OPT_FUNC2, {(void*)opt_height}, "force displayed height", "height" },
3014 { "s", HAS_ARG | OPT_VIDEO, {(void*)opt_frame_size}, "set frame size (WxH or abbreviation)", "size" },
3015 { "fs", OPT_BOOL, {(void*)&is_full_screen}, "force full screen" },
3016 { "an", OPT_BOOL, {(void*)&audio_disable}, "disable audio" },
3017 { "vn", OPT_BOOL, {(void*)&video_disable}, "disable video" },
3018 { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_AUDIO]}, "select desired audio stream", "stream_number" },
3019 { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_VIDEO]}, "select desired video stream", "stream_number" },
3020 { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE]}, "select desired subtitle stream", "stream_number" },
3021 { "ss", HAS_ARG | OPT_FUNC2, {(void*)&opt_seek}, "seek to a given position in seconds", "pos" },
3022 { "t", HAS_ARG | OPT_FUNC2, {(void*)&opt_duration}, "play \"duration\" seconds of audio/video", "duration" },
3023 { "bytes", OPT_INT | HAS_ARG, {(void*)&seek_by_bytes}, "seek by bytes 0=off 1=on -1=auto", "val" },
3024 { "nodisp", OPT_BOOL, {(void*)&display_disable}, "disable graphical display" },
3025 { "f", HAS_ARG, {(void*)opt_format}, "force format", "fmt" },
3026 { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, {(void*)opt_frame_pix_fmt}, "set pixel format", "format" },
3027 { "stats", OPT_BOOL | OPT_EXPERT, {(void*)&show_status}, "show status", "" },
3028 { "debug", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_debug}, "print specific debug info", "" },
3029 { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&workaround_bugs}, "workaround bugs", "" },
3030 { "vismv", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_vismv}, "visualize motion vectors", "" },
3031 { "fast", OPT_BOOL | OPT_EXPERT, {(void*)&fast}, "non spec compliant optimizations", "" },
3032 { "genpts", OPT_BOOL | OPT_EXPERT, {(void*)&genpts}, "generate pts", "" },
3033 { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&decoder_reorder_pts}, "let decoder reorder pts 0=off 1=on -1=auto", ""},
3034 { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&lowres}, "", "" },
3035 { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_loop_filter}, "", "" },
3036 { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_frame}, "", "" },
3037 { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_idct}, "", "" },
3038 { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&idct}, "set idct algo", "algo" },
3039 { "er", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_recognition}, "set error detection threshold (0-4)", "threshold" },
3040 { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_concealment}, "set error concealment options", "bit_mask" },
3041 { "sync", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_sync}, "set audio-video sync. type (type=audio/video/ext)", "type" },
3042 { "threads", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_thread_count}, "thread count", "count" },
3043 { "autoexit", OPT_BOOL | OPT_EXPERT, {(void*)&autoexit}, "exit at the end", "" },
3044 { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&loop}, "set number of times the playback shall be looped", "loop count" },
3045 { "framedrop", OPT_BOOL | OPT_EXPERT, {(void*)&framedrop}, "drop frames when cpu is too slow", "" },
3046 { "window_title", OPT_STRING | HAS_ARG, {(void*)&window_title}, "set window title", "window title" },
3047 #if CONFIG_AVFILTER
3048 { "vf", OPT_STRING | HAS_ARG, {(void*)&vfilters}, "video filters", "filter list" },
3049 #endif
3050 { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, {(void*)&rdftspeed}, "rdft speed", "msecs" },
3051 { "default", OPT_FUNC2 | HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, {(void*)opt_default}, "generic catch all option", "" },
3052 { NULL, },
3053 };
3054
3055 static void show_usage(void)
3056 {
3057 printf("Simple media player\n");
3058 printf("usage: ffplay [options] input_file\n");
3059 printf("\n");
3060 }
3061
3062 static void show_help(void)
3063 {
3064 show_usage();
3065 show_help_options(options, "Main options:\n",
3066 OPT_EXPERT, 0);
3067 show_help_options(options, "\nAdvanced options:\n",
3068 OPT_EXPERT, OPT_EXPERT);
3069 printf("\nWhile playing:\n"
3070 "q, ESC quit\n"
3071 "f toggle full screen\n"
3072 "p, SPC pause\n"
3073 "a cycle audio channel\n"
3074 "v cycle video channel\n"
3075 "t cycle subtitle channel\n"
3076 "w show audio waves\n"
3077 "s activate frame-step mode\n"
3078 "left/right seek backward/forward 10 seconds\n"
3079 "down/up seek backward/forward 1 minute\n"
3080 "mouse click seek to percentage in file corresponding to fraction of width\n"
3081 );
3082 }
3083
3084 static void opt_input_file(const char *filename)
3085 {
3086 if (input_filename) {
3087 fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
3088 filename, input_filename);
3089 exit(1);
3090 }
3091 if (!strcmp(filename, "-"))
3092 filename = "pipe:";
3093 input_filename = filename;
3094 }
3095
3096 /* Called from the main */
3097 int main(int argc, char **argv)
3098 {
3099 int flags, i;
3100
3101 /* register all codecs, demux and protocols */
3102 avcodec_register_all();
3103 #if CONFIG_AVDEVICE
3104 avdevice_register_all();
3105 #endif
3106 #if CONFIG_AVFILTER
3107 avfilter_register_all();
3108 #endif
3109 av_register_all();
3110
3111 for(i=0; i<AVMEDIA_TYPE_NB; i++){
3112 avcodec_opts[i]= avcodec_alloc_context2(i);
3113 }
3114 avformat_opts = avformat_alloc_context();
3115 #if !CONFIG_AVFILTER
3116 sws_opts = sws_getContext(16,16,0, 16,16,0, sws_flags, NULL,NULL,NULL);
3117 #endif
3118
3119 show_banner();
3120
3121 parse_options(argc, argv, options, opt_input_file);
3122
3123 if (!input_filename) {
3124 show_usage();
3125 fprintf(stderr, "An input file must be specified\n");
3126 fprintf(stderr, "Use -h to get full help or, even better, run 'man ffplay'\n");
3127 exit(1);
3128 }
3129
3130 if (display_disable) {
3131 video_disable = 1;
3132 }
3133 flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3134 #if !defined(__MINGW32__) && !defined(__APPLE__)
3135 flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3136 #endif
3137 if (SDL_Init (flags)) {
3138 fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3139 exit(1);
3140 }
3141
3142 if (!display_disable) {
3143 #if HAVE_SDL_VIDEO_SIZE
3144 const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3145 fs_screen_width = vi->current_w;
3146 fs_screen_height = vi->current_h;
3147 #endif
3148 }
3149
3150 SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3151 SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3152 SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3153
3154 av_init_packet(&flush_pkt);
3155 flush_pkt.data= "FLUSH";
3156
3157 cur_stream = stream_open(input_filename, file_iformat);
3158
3159 event_loop();
3160
3161 /* never returns */
3162
3163 return 0;
3164 }