diff options
author | tobigun <tobigun@b956fd51-792f-4845-bead-9b4dfca2ff2c> | 2010-10-27 20:11:05 +0000 |
---|---|---|
committer | tobigun <tobigun@b956fd51-792f-4845-bead-9b4dfca2ff2c> | 2010-10-27 20:11:05 +0000 |
commit | 8021fd92f88d35fe77e9ecac0cc547289ae8c23f (patch) | |
tree | 0461160dfb75fd23b7b8e379b4801aec96178976 /mediaplugin/src | |
parent | 5315ce20f8730c13c3b79f98d90caf8ce74f9cae (diff) | |
download | usdx-8021fd92f88d35fe77e9ecac0cc547289ae8c23f.tar.gz usdx-8021fd92f88d35fe77e9ecac0cc547289ae8c23f.tar.xz usdx-8021fd92f88d35fe77e9ecac0cc547289ae8c23f.zip |
separated video playback and decoder
git-svn-id: svn://svn.code.sf.net/p/ultrastardx/svn/branches/experimental@2691 b956fd51-792f-4845-bead-9b4dfca2ff2c
Diffstat (limited to '')
-rw-r--r-- | mediaplugin/src/media/UVideo.pas | 713 | ||||
-rw-r--r-- | mediaplugin/src/media/UVideoDecoder_FFmpeg.pas | 799 |
2 files changed, 872 insertions, 640 deletions
diff --git a/mediaplugin/src/media/UVideo.pas b/mediaplugin/src/media/UVideo.pas index add7bdc8..6caf9658 100644 --- a/mediaplugin/src/media/UVideo.pas +++ b/mediaplugin/src/media/UVideo.pas @@ -25,12 +25,7 @@ unit UVideo; -{* - * based on 'An ffmpeg and SDL Tutorial' (http://www.dranger.com/ffmpeg/) - *} - // uncomment if you want to see the debug stuff -{.$define DebugDisplay} {.$define DebugFrames} {.$define VideoBenchmark} {.$define Info} @@ -43,30 +38,16 @@ interface {$I switches.inc} -// use BGR-format for accelerated colorspace conversion with swscale -{$IFDEF UseSWScale} - {$DEFINE PIXEL_FMT_BGR} -{$ENDIF} - implementation uses + SDL, SysUtils, Math, - SDL, - avcodec, - avformat, - avutil, - avio, - rational, - {$IFDEF UseSWScale} - swscale, - {$ENDIF} gl, glu, glext, textgl, - UMediaCore_FFmpeg, UCommon, UConfig, ULog, @@ -80,17 +61,14 @@ uses const {$IFDEF PIXEL_FMT_BGR} PIXEL_FMT_OPENGL = GL_BGR; - PIXEL_FMT_FFMPEG = PIX_FMT_BGR24; PIXEL_FMT_SIZE = 3; // looks strange on linux: //PIXEL_FMT_OPENGL = GL_RGBA; - //PIXEL_FMT_FFMPEG = PIX_FMT_BGR32; //PIXEL_FMT_SIZE = 4; {$ELSE} // looks strange on linux: PIXEL_FMT_OPENGL = GL_RGB; - PIXEL_FMT_FFMPEG = PIX_FMT_RGB24; PIXEL_FMT_SIZE = 3; {$ENDIF} @@ -99,35 +77,18 @@ const type IVideo_FFmpeg = interface (IVideo) ['{E640E130-C8C0-4399-AF02-67A3569313AB}'] - function Open(const FileName: IPath): boolean; + function Open(const Decoder: TVideoDecodeStream): boolean; end; TVideo_FFmpeg = class( TInterfacedObject, IVideo_FFmpeg ) private - fOpened: boolean; //**< stream successfully opened + fDecoder: TVideoDecodeStream; fPaused: boolean; //**< stream paused - fEOF: boolean; //**< end-of-file state - - fLoop: boolean; //**< looping enabled - - fStream: PAVStream; - fStreamIndex : integer; - fFormatContext: PAVFormatContext; - fCodecContext: PAVCodecContext; - fCodec: PAVCodec; - fAVFrame: PAVFrame; - fAVFrameRGB: PAVFrame; - - fFrameBuffer: PByte; //**< stores a FFmpeg video frame - fFrameTex: GLuint; //**< OpenGL texture for FrameBuffer - fFrameTexValid: boolean; //**< if true, fFrameTex contains the current frame + fFrameData: PByteArray; + fFrameTex: GLuint; //**< OpenGL texture for FrameBuffer fTexWidth, fTexHeight: cardinal; - {$IFDEF UseSWScale} - fSwScaleContext: PSwsContext; - {$ENDIF} - fScreen: integer; //actual screen to draw on fPosX: double; @@ -141,19 +102,12 @@ type fAlpha: double; fReflectionSpacing: double; - - fAspect: real; //**< width/height ratio fAspectCorrection: TAspectCorrection; - fFrameDuration: extended; //**< duration of a video frame in seconds (= 1/fps) - fFrameTime: extended; //**< video time position (absolute) - fLoopTime: extended; //**< start time of the current loop - fPboEnabled: boolean; fPboId: GLuint; + procedure Reset(); - function DecodeFrame(): boolean; - procedure SynchronizeTime(Frame: PAVFrame; var pts: double); procedure GetVideoRect(var ScreenRect, TexRect: TRectCoords); procedure DrawBorders(ScreenRect: TRectCoords); @@ -165,7 +119,7 @@ type constructor Create; destructor Destroy; override; - function Open(const FileName: IPath): boolean; + function Open(const Decoder: TVideoDecodeStream): boolean; procedure Close; procedure Play; @@ -194,29 +148,26 @@ type * Sub-image of the video frame to draw. * This can be used for zooming or similar purposes. *} - procedure SetFrameRange(Range: TRectCoords); - function GetFrameRange(): TRectCoords; + procedure SetFrameRange(Range: TRectCoords); + function GetFrameRange(): TRectCoords; - function GetFrameAspect(): real; + function GetFrameAspect(): real; - procedure SetAspectCorrection(AspectCorrection: TAspectCorrection); - function GetAspectCorrection(): TAspectCorrection; + procedure SetAspectCorrection(AspectCorrection: TAspectCorrection); + function GetAspectCorrection(): TAspectCorrection; - procedure SetAlpha(Alpha: double); - function GetAlpha(): double; + procedure SetAlpha(Alpha: double); + function GetAlpha(): double; - procedure SetReflectionSpacing(Spacing: double); - function GetReflectionSpacing(): double; + procedure SetReflectionSpacing(Spacing: double); + function GetReflectionSpacing(): double; - procedure GetFrame(Time: Extended); - procedure Draw(); - procedure DrawReflection(); + procedure GetFrame(Time: Extended); + procedure Draw(); + procedure DrawReflection(); end; TVideoPlayback_FFmpeg = class( TInterfacedObject, IVideoPlayback ) - private - fInitialized: boolean; - public function GetName: String; @@ -226,57 +177,18 @@ type function Open(const FileName : IPath): IVideo; end; -var - FFmpegCore: TMediaCore_FFmpeg; - - -// These are called whenever we allocate a frame buffer. -// We use this to store the global_pts in a frame at the time it is allocated. -function PtsGetBuffer(CodecCtx: PAVCodecContext; Frame: PAVFrame): integer; cdecl; -var - pts: Pint64; - VideoPktPts: Pint64; -begin - Result := avcodec_default_get_buffer(CodecCtx, Frame); - VideoPktPts := CodecCtx^.opaque; - if (VideoPktPts <> nil) then - begin - // Note: we must copy the pts instead of passing a pointer, because the packet - // (and with it the pts) might change before a frame is returned by av_decode_video. - pts := av_malloc(sizeof(int64)); - pts^ := VideoPktPts^; - Frame^.opaque := pts; - end; -end; - -procedure PtsReleaseBuffer(CodecCtx: PAVCodecContext; Frame: PAVFrame); cdecl; -begin - if (Frame <> nil) then - av_freep(@Frame^.opaque); - avcodec_default_release_buffer(CodecCtx, Frame); -end; - - {*------------------------------------------------------------------------------ * TVideoPlayback_ffmpeg *------------------------------------------------------------------------------} function TVideoPlayback_FFmpeg.GetName: String; begin - result := 'FFmpeg_Video'; + result := 'OpenGL_VideoPlayback'; end; function TVideoPlayback_FFmpeg.Init(): boolean; begin Result := true; - - if (fInitialized) then - Exit; - fInitialized := true; - - FFmpegCore := TMediaCore_FFmpeg.GetInstance(); - - av_register_all(); end; function TVideoPlayback_FFmpeg.Finalize(): boolean; @@ -287,12 +199,22 @@ end; function TVideoPlayback_FFmpeg.Open(const FileName : IPath): IVideo; var Video: IVideo_FFmpeg; + Decoder: TVideoDecodeStream; begin - Video := TVideo_FFmpeg.Create; - if Video.Open(FileName) then - Result := Video - else - Result := nil; + Result := nil; + + Decoder := VideoDecoder.Open(FileName); + if (Decoder = nil) then + Exit; + + Video := TVideo_FFmpeg.Create(); + if (not Video.Open(Decoder)) then + begin + Decoder.Free; + Exit; + end; + + Result := Video end; @@ -310,165 +232,18 @@ begin glDeleteTextures(1, PGLuint(@fFrameTex)); end; -function TVideo_FFmpeg.Open(const FileName : IPath): boolean; +function TVideo_FFmpeg.Open(const Decoder: TVideoDecodeStream): boolean; var - errnum: Integer; glErr: GLenum; - AudioStreamIndex: integer; begin Result := false; Reset(); - fPboEnabled := PboSupported; - - // use custom 'ufile' protocol for UTF-8 support - errnum := av_open_input_file(fFormatContext, PAnsiChar('ufile:'+FileName.ToUTF8), nil, 0, nil); - if (errnum <> 0) then - begin - Log.LogError('Failed to open file "'+ FileName.ToNative +'" ('+FFmpegCore.GetErrorString(errnum)+')'); - Exit; - end; - - // update video info - if (av_find_stream_info(fFormatContext) < 0) then - begin - Log.LogError('No stream info found', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - Log.LogInfo('VideoStreamIndex : ' + inttostr(fStreamIndex), 'TVideoPlayback_ffmpeg.Open'); - - // find video stream - FFmpegCore.FindStreamIDs(fFormatContext, fStreamIndex, AudioStreamIndex); - if (fStreamIndex < 0) then - begin - Log.LogError('No video stream found', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - - fStream := fFormatContext^.streams[fStreamIndex]; - fCodecContext := fStream^.codec; - - fCodec := avcodec_find_decoder(fCodecContext^.codec_id); - if (fCodec = nil) then - begin - Log.LogError('No matching codec found', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - - // set debug options - fCodecContext^.debug_mv := 0; - fCodecContext^.debug := 0; - - // detect bug-workarounds automatically - fCodecContext^.workaround_bugs := FF_BUG_AUTODETECT; - // error resilience strategy (careful/compliant/agressive/very_aggressive) - //fCodecContext^.error_resilience := FF_ER_CAREFUL; //FF_ER_COMPLIANT; - // allow non spec compliant speedup tricks. - //fCodecContext^.flags2 := fCodecContext^.flags2 or CODEC_FLAG2_FAST; - - // Note: avcodec_open() and avcodec_close() are not thread-safe and will - // fail if called concurrently by different threads. - FFmpegCore.LockAVCodec(); - try - errnum := avcodec_open(fCodecContext, fCodec); - finally - FFmpegCore.UnlockAVCodec(); - end; - if (errnum < 0) then - begin - Log.LogError('No matching codec found', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - - // register custom callbacks for pts-determination - fCodecContext^.get_buffer := PtsGetBuffer; - fCodecContext^.release_buffer := PtsReleaseBuffer; - - {$ifdef DebugDisplay} - DebugWriteln('Found a matching Codec: '+ fCodecContext^.Codec.Name + sLineBreak + - sLineBreak + - ' Width = '+inttostr(fCodecContext^.width) + - ', Height='+inttostr(fCodecContext^.height) + sLineBreak + - ' Aspect : '+inttostr(fCodecContext^.sample_aspect_ratio.num) + '/' + - inttostr(fCodecContext^.sample_aspect_ratio.den) + sLineBreak + - ' Framerate : '+inttostr(fCodecContext^.time_base.num) + '/' + - inttostr(fCodecContext^.time_base.den)); - {$endif} - - // allocate space for decoded frame and rgb frame - fAVFrame := avcodec_alloc_frame(); - fAVFrameRGB := avcodec_alloc_frame(); - fFrameBuffer := av_malloc(avpicture_get_size(PIXEL_FMT_FFMPEG, - fCodecContext^.width, fCodecContext^.height)); - - if ((fAVFrame = nil) or (fAVFrameRGB = nil) or (fFrameBuffer = nil)) then - begin - Log.LogError('Failed to allocate buffers', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - - // TODO: pad data for OpenGL to GL_UNPACK_ALIGNMENT - // (otherwise video will be distorted if width/height is not a multiple of the alignment) - errnum := avpicture_fill(PAVPicture(fAVFrameRGB), fFrameBuffer, PIXEL_FMT_FFMPEG, - fCodecContext^.width, fCodecContext^.height); - if (errnum < 0) then - begin - Log.LogError('avpicture_fill failed: ' + FFmpegCore.GetErrorString(errnum), 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - - // calculate some information for video display - fAspect := av_q2d(fCodecContext^.sample_aspect_ratio); - if (fAspect = 0) then - fAspect := fCodecContext^.width / - fCodecContext^.height - else - fAspect := fAspect * fCodecContext^.width / - fCodecContext^.height; - - fFrameDuration := 1/av_q2d(fStream^.r_frame_rate); - - // hack to get reasonable framerate (for divx and others) - if (fFrameDuration < 0.02) then // 0.02 <-> 50 fps - begin - fFrameDuration := av_q2d(fStream^.r_frame_rate); - while (fFrameDuration > 50) do - fFrameDuration := fFrameDuration/10; - fFrameDuration := 1/fFrameDuration; - end; - - Log.LogInfo('Framerate: '+inttostr(floor(1/fFrameDuration))+'fps', 'TVideoPlayback_ffmpeg.Open'); - - {$IFDEF UseSWScale} - // if available get a SWScale-context -> faster than the deprecated img_convert(). - // SWScale has accelerated support for PIX_FMT_RGB32/PIX_FMT_BGR24/PIX_FMT_BGR565/PIX_FMT_BGR555. - // Note: PIX_FMT_RGB32 is a BGR- and not an RGB-format (maybe a bug)!!! - // The BGR565-formats (GL_UNSIGNED_SHORT_5_6_5) is way too slow because of its - // bad OpenGL support. The BGR formats have MMX(2) implementations but no speed-up - // could be observed in comparison to the RGB versions. - fSwScaleContext := sws_getContext( - fCodecContext^.width, fCodecContext^.height, - fCodecContext^.pix_fmt, - fCodecContext^.width, fCodecContext^.height, - PIXEL_FMT_FFMPEG, - SWS_FAST_BILINEAR, nil, nil, nil); - if (fSwScaleContext = nil) then - begin - Log.LogError('Failed to get swscale context', 'TVideoPlayback_ffmpeg.Open'); - Close(); - Exit; - end; - {$ENDIF} - - fTexWidth := Round(Power(2, Ceil(Log2(fCodecContext^.width)))); - fTexHeight := Round(Power(2, Ceil(Log2(fCodecContext^.height)))); + fDecoder := Decoder; + fTexWidth := Round(Power(2, Ceil(Log2(Decoder.GetFrameWidth())))); + fTexHeight := Round(Power(2, Ceil(Log2(Decoder.GetFrameHeight())))); + fPboEnabled := PboSupported; if (fPboEnabled) then begin glGetError(); @@ -477,7 +252,7 @@ begin glBindBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB, fPboId); glBufferDataARB( GL_PIXEL_UNPACK_BUFFER_ARB, - fCodecContext^.width * fCodecContext^.height * PIXEL_FMT_SIZE, + Decoder.GetFrameWidth() * Decoder.GetFrameHeight() * PIXEL_FMT_SIZE, nil, GL_STREAM_DRAW_ARB); glBindBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB, 0); @@ -498,7 +273,6 @@ begin glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); - fOpened := true; Result := true; end; @@ -507,18 +281,7 @@ begin // close previously opened video Close(); - fOpened := False; fPaused := False; - fFrameDuration := 0; - fFrameTime := 0; - fStream := nil; - fStreamIndex := -1; - fFrameTexValid := false; - - fEOF := false; - - fLoop := false; - fLoopTime := 0; fPboId := 0; @@ -541,319 +304,33 @@ begin fReflectionSpacing := 0; end; + procedure TVideo_FFmpeg.Close; begin - if (fFrameBuffer <> nil) then - av_free(fFrameBuffer); - if (fAVFrameRGB <> nil) then - av_free(fAVFrameRGB); - if (fAVFrame <> nil) then - av_free(fAVFrame); - - fAVFrame := nil; - fAVFrameRGB := nil; - fFrameBuffer := nil; - - if (fCodecContext <> nil) then + if (fDecoder <> nil) then begin - // avcodec_close() is not thread-safe - FFmpegCore.LockAVCodec(); - try - avcodec_close(fCodecContext); - finally - FFmpegCore.UnlockAVCodec(); - end; + fDecoder.close(); + FreeAndNil(fDecoder); end; - if (fFormatContext <> nil) then - av_close_input_file(fFormatContext); - - fCodecContext := nil; - fFormatContext := nil; - if (fPboId <> 0) then glDeleteBuffersARB(1, @fPboId); - - fOpened := False; -end; - -procedure TVideo_FFmpeg.SynchronizeTime(Frame: PAVFrame; var pts: double); -var - FrameDelay: double; -begin - if (pts <> 0) then - begin - // if we have pts, set video clock to it - fFrameTime := pts; - end else - begin - // if we aren't given a pts, set it to the clock - pts := fFrameTime; - end; - // update the video clock - FrameDelay := av_q2d(fCodecContext^.time_base); - // if we are repeating a frame, adjust clock accordingly - FrameDelay := FrameDelay + Frame^.repeat_pict * (FrameDelay * 0.5); - fFrameTime := fFrameTime + FrameDelay; -end; - -{** - * Decode a new frame from the video stream. - * The decoded frame is stored in fAVFrame. fFrameTime is updated to the new frame's - * time. - * @param pts will be updated to the presentation time of the decoded frame. - * returns true if a frame could be decoded. False if an error or EOF occured. - *} -function TVideo_FFmpeg.DecodeFrame(): boolean; -var - FrameFinished: Integer; - VideoPktPts: int64; - pbIOCtx: PByteIOContext; - errnum: integer; - AVPacket: TAVPacket; - pts: double; -begin - Result := false; - FrameFinished := 0; - - if fEOF then - Exit; - - // read packets until we have a finished frame (or there are no more packets) - while (FrameFinished = 0) do - begin - errnum := av_read_frame(fFormatContext, AVPacket); - if (errnum < 0) then - begin - // failed to read a frame, check reason - - {$IF (LIBAVFORMAT_VERSION_MAJOR >= 52)} - pbIOCtx := fFormatContext^.pb; - {$ELSE} - pbIOCtx := @fFormatContext^.pb; - {$IFEND} - - // check for end-of-file (EOF is not an error) - if (url_feof(pbIOCtx) <> 0) then - begin - fEOF := true; - Exit; - end; - - // check for errors - if (url_ferror(pbIOCtx) <> 0) then - begin - Log.LogError('Video decoding file error', 'TVideoPlayback_FFmpeg.DecodeFrame'); - Exit; - end; - - // url_feof() does not detect an EOF for some mov-files (e.g. deluxe.mov) - // so we have to do it this way. - if ((fFormatContext^.file_size <> 0) and - (pbIOCtx^.pos >= fFormatContext^.file_size)) then - begin - fEOF := true; - Exit; - end; - - // error occured, log and exit - Log.LogError('Video decoding error', 'TVideoPlayback_FFmpeg.DecodeFrame'); - Exit; - end; - - // if we got a packet from the video stream, then decode it - if (AVPacket.stream_index = fStreamIndex) then - begin - // save pts to be stored in pFrame in first call of PtsGetBuffer() - VideoPktPts := AVPacket.pts; - fCodecContext^.opaque := @VideoPktPts; - - // decode packet - avcodec_decode_video(fCodecContext, fAVFrame, - frameFinished, AVPacket.data, AVPacket.size); - - // reset opaque data - fCodecContext^.opaque := nil; - - // update pts - if (AVPacket.dts <> AV_NOPTS_VALUE) then - begin - pts := AVPacket.dts; - end - else if ((fAVFrame^.opaque <> nil) and - (Pint64(fAVFrame^.opaque)^ <> AV_NOPTS_VALUE)) then - begin - pts := Pint64(fAVFrame^.opaque)^; - end - else - begin - pts := 0; - end; - - if fStream^.start_time <> AV_NOPTS_VALUE then - pts := pts - fStream^.start_time; - - pts := pts * av_q2d(fStream^.time_base); - - // synchronize time on each complete frame - if (frameFinished <> 0) then - SynchronizeTime(fAVFrame, pts); - end; - - // free the packet from av_read_frame - av_free_packet( @AVPacket ); - end; - - Result := true; end; procedure TVideo_FFmpeg.GetFrame(Time: Extended); var - errnum: Integer; glErr: GLenum; - CurrentTime: Extended; - TimeDiff: Extended; - DropFrameCount: Integer; - i: Integer; - Success: boolean; BufferPtr: PGLvoid; -const - SKIP_FRAME_DIFF = 0.010; // start skipping if we are >= 10ms too late begin - if not fOpened then + if (fDecoder = nil) then Exit; if fPaused then Exit; - {* - * Synchronization - begin - *} - - // requested stream position (relative to the last loop's start) - if (fLoop) then - CurrentTime := Time - fLoopTime - else - CurrentTime := Time; - - // check if current texture still contains the active frame - if (fFrameTexValid) then - begin - // time since the last frame was returned - TimeDiff := CurrentTime - fFrameTime; - - {$IFDEF DebugDisplay} - DebugWriteln('Time: '+inttostr(floor(Time*1000)) + sLineBreak + - 'VideoTime: '+inttostr(floor(fFrameTime*1000)) + sLineBreak + - 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + - 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); - {$endif} - - // check if time has reached the next frame - if (TimeDiff < fFrameDuration) then - begin - {$ifdef DebugFrames} - // frame delay debug display - GoldenRec.Spawn(200,15,1,16,0,-1,ColoredStar,$00ff00); - {$endif} - - {$IFDEF DebugDisplay} - DebugWriteln('not getting new frame' + sLineBreak + - 'Time: '+inttostr(floor(Time*1000)) + sLineBreak + - 'VideoTime: '+inttostr(floor(fFrameTime*1000)) + sLineBreak + - 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + - 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); - {$endif} - - // we do not need a new frame now - Exit; - end; - end; - - {$IFDEF VideoBenchmark} - Log.BenchmarkStart(15); - {$ENDIF} - - // fetch new frame (updates fFrameTime) - Success := DecodeFrame(); - TimeDiff := CurrentTime - fFrameTime; - - // check if we have to skip frames - // Either if we are one frame behind or if the skip threshold has been reached. - // Do not skip if the difference is less than fFrameDuration as there is no next frame. - // Note: We assume that fFrameDuration is the length of one frame. - if (TimeDiff >= Max(fFrameDuration, SKIP_FRAME_DIFF)) then - begin - {$IFDEF DebugFrames} - //frame drop debug display - GoldenRec.Spawn(200,55,1,16,0,-1,ColoredStar,$ff0000); - {$ENDIF} - {$IFDEF DebugDisplay} - DebugWriteln('skipping frames' + sLineBreak + - 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + - 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); - {$endif} - - // update video-time - DropFrameCount := Trunc(TimeDiff / fFrameDuration); - fFrameTime := fFrameTime + DropFrameCount*fFrameDuration; - - // skip frames - for i := 1 to DropFrameCount do - Success := DecodeFrame(); - end; - - // check if we got an EOF or error - if (not Success) then - begin - if fLoop then - begin - // we have to loop, so rewind - SetPosition(0); - // record the start-time of the current loop, so we can - // determine the position in the stream (fFrameTime-fLoopTime) later. - fLoopTime := Time; - end; - Exit; - end; - - {* - * Synchronization - end - *} - - // TODO: support for pan&scan - //if (fAVFrame.pan_scan <> nil) then - //begin - // Writeln(Format('PanScan: %d/%d', [fAVFrame.pan_scan.width, fAVFrame.pan_scan.height])); - //end; - - // otherwise we convert the pixeldata from YUV to RGB - {$IFDEF UseSWScale} - errnum := sws_scale(fSwScaleContext, @fAVFrame.data, @fAVFrame.linesize, - 0, fCodecContext^.Height, - @fAVFrameRGB.data, @fAVFrameRGB.linesize); - {$ELSE} - // img_convert from lib/ffmpeg/avcodec.pas is actually deprecated. - // If ./configure does not find SWScale then this gives the error - // that the identifier img_convert is not known or similar. - // I think this should be removed, but am not sure whether there should - // be some other replacement or a warning, Therefore, I leave it for now. - // April 2009, mischi - errnum := img_convert(PAVPicture(fAVFrameRGB), PIXEL_FMT_FFMPEG, - PAVPicture(fAVFrame), fCodecContext^.pix_fmt, - fCodecContext^.width, fCodecContext^.height); - {$ENDIF} - - if (errnum < 0) then - begin - Log.LogError('Image conversion failed', 'TVideoPlayback_ffmpeg.GetFrame'); + fFrameData := fDecoder.GetFrame(Time); + if (fFrameData = nil) then Exit; - end; - - {$IFDEF VideoBenchmark} - Log.BenchmarkEnd(15); - Log.BenchmarkStart(16); - {$ENDIF} // TODO: data is not padded, so we will need to tell OpenGL. // Or should we add padding with avpicture_fill? (check which one is faster) @@ -866,8 +343,8 @@ begin begin glBindTexture(GL_TEXTURE_2D, fFrameTex); glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, - fCodecContext^.width, fCodecContext^.height, - PIXEL_FMT_OPENGL, GL_UNSIGNED_BYTE, fAVFrameRGB^.data[0]); + fDecoder.GetFrameWidth(), fDecoder.GetFrameHeight(), + PIXEL_FMT_OPENGL, GL_UNSIGNED_BYTE, fFrameData); end else // fPboEnabled begin @@ -875,15 +352,15 @@ begin glBindBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB, fPboId); glBufferDataARB(GL_PIXEL_UNPACK_BUFFER_ARB, - fCodecContext^.height * fCodecContext^.width * PIXEL_FMT_SIZE, + fDecoder.GetFrameHeight() * fDecoder.GetFrameWidth() * PIXEL_FMT_SIZE, nil, GL_STREAM_DRAW_ARB); bufferPtr := glMapBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB, GL_WRITE_ONLY_ARB); if(bufferPtr <> nil) then begin - Move(fAVFrameRGB^.data[0]^, bufferPtr^, - fCodecContext^.height * fCodecContext^.width * PIXEL_FMT_SIZE); + Move(fFrameData[0], bufferPtr^, + fDecoder.GetFrameHeight() * fDecoder.GetFrameWidth() * PIXEL_FMT_SIZE); // release pointer to mapping buffer glUnmapBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB); @@ -891,7 +368,7 @@ begin glBindTexture(GL_TEXTURE_2D, fFrameTex); glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, - fCodecContext^.width, fCodecContext^.height, + fDecoder.GetFrameWidth(), fDecoder.GetFrameHeight(), PIXEL_FMT_OPENGL, GL_UNSIGNED_BYTE, nil); glBindBufferARB(GL_PIXEL_UNPACK_BUFFER_ARB, 0); @@ -905,9 +382,6 @@ begin // reset to default glTexEnvi(GL_TEXTURE_ENV, GL_TEXTURE_ENV_MODE, GL_MODULATE); - if (not fFrameTexValid) then - fFrameTexValid := true; - {$ifdef DebugFrames} //frame decode debug display GoldenRec.Spawn(200, 35, 1, 16, 0, -1, ColoredStar, $ffff00); @@ -924,13 +398,14 @@ procedure TVideo_FFmpeg.GetVideoRect(var ScreenRect, TexRect: TRectCoords); var ScreenAspect: double; // aspect of screen resolution ScaledVideoWidth, ScaledVideoHeight: double; - + FrameAspect: real; begin // Three aspects to take into account: // 1. Screen/display resolution (e.g. 1920x1080 -> 16:9) // 2. Render aspect (fWidth x fHeight -> variable) // 3. Movie aspect (video frame aspect stored in fAspect) ScreenAspect := fWidth*((ScreenW/Screens)/RenderW)/(fHeight*(ScreenH/RenderH)); + FrameAspect := fDecoder.GetFrameAspect(); case fAspectCorrection of acoStretch: begin @@ -939,26 +414,26 @@ begin end; acoCrop: begin - if (ScreenAspect >= fAspect) then + if (ScreenAspect >= FrameAspect) then begin ScaledVideoWidth := fWidth; - ScaledVideoHeight := fHeight * ScreenAspect/fAspect; + ScaledVideoHeight := fHeight * ScreenAspect/FrameAspect; end else begin ScaledVideoHeight := fHeight; - ScaledVideoWidth := fWidth * fAspect/ScreenAspect; + ScaledVideoWidth := fWidth * FrameAspect/ScreenAspect; end; end; acoLetterBox: begin - if (ScreenAspect <= fAspect) then + if (ScreenAspect <= FrameAspect) then begin ScaledVideoWidth := fWidth; - ScaledVideoHeight := fHeight * ScreenAspect/fAspect; + ScaledVideoHeight := fHeight * ScreenAspect/FrameAspect; end else begin ScaledVideoHeight := fHeight; - ScaledVideoWidth := fWidth * fAspect/ScreenAspect; + ScaledVideoWidth := fWidth * FrameAspect/ScreenAspect; end; end else raise Exception.Create('Unhandled aspect correction!'); @@ -972,10 +447,10 @@ begin // texture contains right/lower (power-of-2) padding. // Determine the texture coords of the video frame. - TexRect.Left := (fCodecContext^.width / fTexWidth) * fFrameRange.Left; - TexRect.Right := (fCodecContext^.width / fTexWidth) * fFrameRange.Right; - TexRect.Upper := (fCodecContext^.height / fTexHeight) * fFrameRange.Upper; - TexRect.Lower := (fCodecContext^.height / fTexHeight) * fFrameRange.Lower; + TexRect.Left := (fDecoder.GetFrameWidth() / fTexWidth) * fFrameRange.Left; + TexRect.Right := (fDecoder.GetFrameWidth() / fTexWidth) * fFrameRange.Right; + TexRect.Upper := (fDecoder.GetFrameHeight() / fTexHeight) * fFrameRange.Upper; + TexRect.Lower := (fDecoder.GetFrameHeight() / fTexHeight) * fFrameRange.Lower; end; procedure TVideo_FFmpeg.DrawBorders(ScreenRect: TRectCoords); @@ -1061,7 +536,7 @@ var begin // exit if there's nothing to draw - if (not fOpened) then + if (fDecoder = nil) then Exit; {$IFDEF VideoBenchmark} @@ -1140,7 +615,7 @@ var begin // exit if there's nothing to draw - if (not fOpened) then + if (fDecoder = nil) then Exit; // get texture and screen positions @@ -1265,64 +740,22 @@ end; procedure TVideo_FFmpeg.SetLoop(Enable: boolean); begin - fLoop := Enable; - fLoopTime := 0; + fDecoder.SetLoop(Enable); end; function TVideo_FFmpeg.GetLoop(): boolean; begin - Result := fLoop; + Result := fDecoder.GetLoop(); end; -{** - * Sets the stream's position. - * The stream is set to the first keyframe with timestamp <= Time. - * Note that fFrameTime is set to Time no matter if the actual position seeked to is - * at Time or the time of a preceding keyframe. fFrameTime will be updated to the - * actual frame time when GetFrame() is called the next time. - * @param Time new position in seconds - *} procedure TVideo_FFmpeg.SetPosition(Time: real); -var - SeekFlags: integer; begin - if not fOpened then - Exit; - - if (Time < 0) then - Time := 0; - - // TODO: handle fLoop-times - //Time := Time mod VideoDuration; - - // Do not use the AVSEEK_FLAG_ANY here. It will seek to any frame, even - // non keyframes (P-/B-frames). It will produce corrupted video frames as - // FFmpeg does not use the information of the preceding I-frame. - // The picture might be gray or green until the next keyframe occurs. - // Instead seek the first keyframe smaller than the requested time - // (AVSEEK_FLAG_BACKWARD). As this can be some seconds earlier than the - // requested time, let the sync in GetFrame() do its job. - SeekFlags := AVSEEK_FLAG_BACKWARD; - - fFrameTime := Time; - fEOF := false; - fFrameTexValid := false; - - if (av_seek_frame(fFormatContext, - fStreamIndex, - Round(Time / av_q2d(fStream^.time_base)), - SeekFlags) < 0) then - begin - Log.LogError('av_seek_frame() failed', 'TVideoPlayback_ffmpeg.SetPosition'); - Exit; - end; - - avcodec_flush_buffers(fCodecContext); + fDecoder.SetPosition(Time); end; function TVideo_FFmpeg.GetPosition: real; begin - Result := fFrameTime; + Result := fDecoder.GetPosition(); end; procedure TVideo_FFmpeg.SetScreen(Screen: integer); @@ -1386,7 +819,7 @@ end; function TVideo_FFmpeg.GetFrameAspect(): real; begin - Result := fAspect; + Result := fDecoder.GetFrameAspect(); end; diff --git a/mediaplugin/src/media/UVideoDecoder_FFmpeg.pas b/mediaplugin/src/media/UVideoDecoder_FFmpeg.pas new file mode 100644 index 00000000..6b6dd901 --- /dev/null +++ b/mediaplugin/src/media/UVideoDecoder_FFmpeg.pas @@ -0,0 +1,799 @@ +{* UltraStar Deluxe - Karaoke Game + * + * UltraStar Deluxe is the legal property of its developers, whose names + * are too numerous to list here. Please refer to the COPYRIGHT + * file distributed with this source distribution. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; see the file COPYING. If not, write to + * the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, + * Boston, MA 02110-1301, USA. + * + * $URL$ + * $Id$ + *} + +unit UVideoDecoder_FFmpeg; + +{* + * based on 'An ffmpeg and SDL Tutorial' (http://www.dranger.com/ffmpeg/) + *} + +// uncomment if you want to see the debug stuff +{.$define DebugDisplay} +{.$define DebugFrames} + +interface + +{$IFDEF FPC} + {$MODE Delphi} +{$ENDIF} + +{$I switches.inc} + +// use BGR-format for accelerated colorspace conversion with swscale +{$IFDEF UseSWScale} + {$DEFINE PIXEL_FMT_BGR} +{$ENDIF} + +implementation + +uses + SysUtils, + Math, + avcodec, + avformat, + avutil, + avio, + rational, + {$IFDEF UseSWScale} + swscale, + {$ENDIF} + UMediaCore_FFmpeg, + UCommon, + UConfig, + ULog, + UMusic, + UPath; + +{$DEFINE PIXEL_FMT_BGR} + +const +{$IFDEF PIXEL_FMT_BGR} + PIXEL_FMT_FFMPEG = PIX_FMT_BGR24; + PIXEL_FMT_SIZE = 3; + + // looks strange on linux: + //PIXEL_FMT_FFMPEG = PIX_FMT_BGR32; + //PIXEL_FMT_SIZE = 4; +{$ELSE} + // looks strange on linux: + PIXEL_FMT_FFMPEG = PIX_FMT_RGB24; + PIXEL_FMT_SIZE = 3; +{$ENDIF} + +type + TVideoDecodeStream_FFmpeg = class (TVideoDecodeStream) + private + fOpened: boolean; //**< stream successfully opened + fEOF: boolean; //**< end-of-file state + + fLoop: boolean; //**< looping enabled + + fStream: PAVStream; + fStreamIndex : integer; + fFormatContext: PAVFormatContext; + fCodecContext: PAVCodecContext; + fCodec: PAVCodec; + + fAVFrame: PAVFrame; + fAVFrameRGB: PAVFrame; + + fFrameBuffer: PByte; //**< stores a FFmpeg video frame + fFrameTexValid: boolean; //**< if true, fFrameTex contains the current frame + + {$IFDEF UseSWScale} + fSwScaleContext: PSwsContext; + {$ENDIF} + + fAspect: real; //**< width/height ratio + + fFrameDuration: extended; //**< duration of a video frame in seconds (= 1/fps) + fFrameTime: extended; //**< video time position (absolute) + fLoopTime: extended; //**< start time of the current loop + + procedure Reset(); + function DecodeFrame(): boolean; + procedure SynchronizeTime(Frame: PAVFrame; var pts: double); + + public + constructor Create; + destructor Destroy; override; + + function Open(const FileName: IPath): boolean; override; + procedure Close; override; + + procedure SetLoop(Enable: boolean); override; + function GetLoop(): boolean; override; + + procedure SetPosition(Time: real); override; + function GetPosition: real; override; + + function GetFrameWidth(): integer; override; + function GetFrameHeight(): integer; override; + + function GetFrameAspect(): real; override; + function GetFrame(Time: Extended): PByteArray; override; + end; + + TVideoDecoder_FFmpeg = class( TInterfacedObject, IVideoDecoder ) + private + fInitialized: boolean; + + public + function GetName: String; + + function InitializeDecoder(): boolean; + function FinalizeDecoder: boolean; + + function Open(const FileName: IPath): TVideoDecodeStream; + end; + +var + FFmpegCore: TMediaCore_FFmpeg; + + +// These are called whenever we allocate a frame buffer. +// We use this to store the global_pts in a frame at the time it is allocated. +function PtsGetBuffer(CodecCtx: PAVCodecContext; Frame: PAVFrame): integer; cdecl; +var + pts: Pint64; + VideoPktPts: Pint64; +begin + Result := avcodec_default_get_buffer(CodecCtx, Frame); + VideoPktPts := CodecCtx^.opaque; + if (VideoPktPts <> nil) then + begin + // Note: we must copy the pts instead of passing a pointer, because the packet + // (and with it the pts) might change before a frame is returned by av_decode_video. + pts := av_malloc(sizeof(int64)); + pts^ := VideoPktPts^; + Frame^.opaque := pts; + end; +end; + +procedure PtsReleaseBuffer(CodecCtx: PAVCodecContext; Frame: PAVFrame); cdecl; +begin + if (Frame <> nil) then + av_freep(@Frame^.opaque); + avcodec_default_release_buffer(CodecCtx, Frame); +end; + + +{*------------------------------------------------------------------------------ + * TVideoPlayback_ffmpeg + *------------------------------------------------------------------------------} + +function TVideoDecoder_FFmpeg.GetName: String; +begin + result := 'FFmpeg_VideoDecoder'; +end; + +function TVideoDecoder_FFmpeg.InitializeDecoder(): boolean; +begin + Result := true; + + if (fInitialized) then + Exit; + fInitialized := true; + + FFmpegCore := TMediaCore_FFmpeg.GetInstance(); + + av_register_all(); +end; + +function TVideoDecoder_FFmpeg.FinalizeDecoder(): boolean; +begin + Result := true; +end; + +function TVideoDecoder_FFmpeg.Open(const FileName : IPath): TVideoDecodeStream; +var + Stream: TVideoDecodeStream_FFmpeg; +begin + Result := nil; + + Stream := TVideoDecodeStream_FFmpeg.Create; + if (not Stream.Open(FileName)) then + begin + Stream.Free; + Exit; + end; + + Result := Stream +end; + + +{* TVideoDecoder_FFmpeg *} + +constructor TVideoDecodeStream_FFmpeg.Create; +begin + Reset(); +end; + +destructor TVideoDecodeStream_FFmpeg.Destroy; +begin + Close(); +end; + +function TVideoDecodeStream_FFmpeg.Open(const FileName: IPath): boolean; +var + errnum: Integer; + AudioStreamIndex: integer; +begin + Result := false; + Reset(); + + // use custom 'ufile' protocol for UTF-8 support + errnum := av_open_input_file(fFormatContext, PAnsiChar('ufile:'+FileName.ToUTF8), nil, 0, nil); + if (errnum <> 0) then + begin + Log.LogError('Failed to open file "'+ FileName.ToNative +'" ('+FFmpegCore.GetErrorString(errnum)+')'); + Exit; + end; + + // update video info + if (av_find_stream_info(fFormatContext) < 0) then + begin + Log.LogError('No stream info found', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + Log.LogInfo('VideoStreamIndex : ' + inttostr(fStreamIndex), 'TVideoPlayback_ffmpeg.Open'); + + // find video stream + FFmpegCore.FindStreamIDs(fFormatContext, fStreamIndex, AudioStreamIndex); + if (fStreamIndex < 0) then + begin + Log.LogError('No video stream found', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + + fStream := fFormatContext^.streams[fStreamIndex]; + fCodecContext := fStream^.codec; + + fCodec := avcodec_find_decoder(fCodecContext^.codec_id); + if (fCodec = nil) then + begin + Log.LogError('No matching codec found', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + + // set debug options + fCodecContext^.debug_mv := 0; + fCodecContext^.debug := 0; + + // detect bug-workarounds automatically + fCodecContext^.workaround_bugs := FF_BUG_AUTODETECT; + // error resilience strategy (careful/compliant/agressive/very_aggressive) + //fCodecContext^.error_resilience := FF_ER_CAREFUL; //FF_ER_COMPLIANT; + // allow non spec compliant speedup tricks. + //fCodecContext^.flags2 := fCodecContext^.flags2 or CODEC_FLAG2_FAST; + + // Note: avcodec_open() and avcodec_close() are not thread-safe and will + // fail if called concurrently by different threads. + FFmpegCore.LockAVCodec(); + try + errnum := avcodec_open(fCodecContext, fCodec); + finally + FFmpegCore.UnlockAVCodec(); + end; + if (errnum < 0) then + begin + Log.LogError('No matching codec found', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + + // register custom callbacks for pts-determination + fCodecContext^.get_buffer := PtsGetBuffer; + fCodecContext^.release_buffer := PtsReleaseBuffer; + + {$ifdef DebugDisplay} + DebugWriteln('Found a matching Codec: '+ fCodecContext^.Codec.Name + sLineBreak + + sLineBreak + + ' Width = '+inttostr(fCodecContext^.width) + + ', Height='+inttostr(fCodecContext^.height) + sLineBreak + + ' Aspect : '+inttostr(fCodecContext^.sample_aspect_ratio.num) + '/' + + inttostr(fCodecContext^.sample_aspect_ratio.den) + sLineBreak + + ' Framerate : '+inttostr(fCodecContext^.time_base.num) + '/' + + inttostr(fCodecContext^.time_base.den)); + {$endif} + + // allocate space for decoded frame and rgb frame + fAVFrame := avcodec_alloc_frame(); + fAVFrameRGB := avcodec_alloc_frame(); + fFrameBuffer := av_malloc(avpicture_get_size(PIXEL_FMT_FFMPEG, + fCodecContext^.width, fCodecContext^.height)); + + if ((fAVFrame = nil) or (fAVFrameRGB = nil) or (fFrameBuffer = nil)) then + begin + Log.LogError('Failed to allocate buffers', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + + // TODO: pad data for OpenGL to GL_UNPACK_ALIGNMENT + // (otherwise video will be distorted if width/height is not a multiple of the alignment) + errnum := avpicture_fill(PAVPicture(fAVFrameRGB), fFrameBuffer, PIXEL_FMT_FFMPEG, + fCodecContext^.width, fCodecContext^.height); + if (errnum < 0) then + begin + Log.LogError('avpicture_fill failed: ' + FFmpegCore.GetErrorString(errnum), 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + + // calculate some information for video display + fAspect := av_q2d(fCodecContext^.sample_aspect_ratio); + if (fAspect = 0) then + fAspect := fCodecContext^.width / + fCodecContext^.height + else + fAspect := fAspect * fCodecContext^.width / + fCodecContext^.height; + + fFrameDuration := 1/av_q2d(fStream^.r_frame_rate); + + // hack to get reasonable framerate (for divx and others) + if (fFrameDuration < 0.02) then // 0.02 <-> 50 fps + begin + fFrameDuration := av_q2d(fStream^.r_frame_rate); + while (fFrameDuration > 50) do + fFrameDuration := fFrameDuration/10; + fFrameDuration := 1/fFrameDuration; + end; + + Log.LogInfo('Framerate: '+inttostr(floor(1/fFrameDuration))+'fps', 'TVideoPlayback_ffmpeg.Open'); + + {$IFDEF UseSWScale} + // if available get a SWScale-context -> faster than the deprecated img_convert(). + // SWScale has accelerated support for PIX_FMT_RGB32/PIX_FMT_BGR24/PIX_FMT_BGR565/PIX_FMT_BGR555. + // Note: PIX_FMT_RGB32 is a BGR- and not an RGB-format (maybe a bug)!!! + // The BGR565-formats (GL_UNSIGNED_SHORT_5_6_5) is way too slow because of its + // bad OpenGL support. The BGR formats have MMX(2) implementations but no speed-up + // could be observed in comparison to the RGB versions. + fSwScaleContext := sws_getContext( + fCodecContext^.width, fCodecContext^.height, + fCodecContext^.pix_fmt, + fCodecContext^.width, fCodecContext^.height, + PIXEL_FMT_FFMPEG, + SWS_FAST_BILINEAR, nil, nil, nil); + if (fSwScaleContext = nil) then + begin + Log.LogError('Failed to get swscale context', 'TVideoPlayback_ffmpeg.Open'); + Close(); + Exit; + end; + {$ENDIF} + + fOpened := true; + Result := true; +end; + +procedure TVideoDecodeStream_FFmpeg.Reset(); +begin + // close previously opened video + Close(); + + fOpened := False; + fFrameDuration := 0; + fFrameTime := 0; + fStream := nil; + fStreamIndex := -1; + fFrameTexValid := false; + + fEOF := false; + + fLoop := false; + fLoopTime := 0; +end; + +procedure TVideoDecodeStream_FFmpeg.Close; +begin + if (fFrameBuffer <> nil) then + av_free(fFrameBuffer); + if (fAVFrameRGB <> nil) then + av_free(fAVFrameRGB); + if (fAVFrame <> nil) then + av_free(fAVFrame); + + fAVFrame := nil; + fAVFrameRGB := nil; + fFrameBuffer := nil; + + if (fCodecContext <> nil) then + begin + // avcodec_close() is not thread-safe + FFmpegCore.LockAVCodec(); + try + avcodec_close(fCodecContext); + finally + FFmpegCore.UnlockAVCodec(); + end; + end; + + if (fFormatContext <> nil) then + av_close_input_file(fFormatContext); + + fCodecContext := nil; + fFormatContext := nil; + + fOpened := False; +end; + +procedure TVideoDecodeStream_FFmpeg.SynchronizeTime(Frame: PAVFrame; var pts: double); +var + FrameDelay: double; +begin + if (pts <> 0) then + begin + // if we have pts, set video clock to it + fFrameTime := pts; + end else + begin + // if we aren't given a pts, set it to the clock + pts := fFrameTime; + end; + // update the video clock + FrameDelay := av_q2d(fCodecContext^.time_base); + // if we are repeating a frame, adjust clock accordingly + FrameDelay := FrameDelay + Frame^.repeat_pict * (FrameDelay * 0.5); + fFrameTime := fFrameTime + FrameDelay; +end; + +{** + * Decode a new frame from the video stream. + * The decoded frame is stored in fAVFrame. fFrameTime is updated to the new frame's + * time. + * @param pts will be updated to the presentation time of the decoded frame. + * returns true if a frame could be decoded. False if an error or EOF occured. + *} +function TVideoDecodeStream_FFmpeg.DecodeFrame(): boolean; +var + FrameFinished: Integer; + VideoPktPts: int64; + pbIOCtx: PByteIOContext; + errnum: integer; + AVPacket: TAVPacket; + pts: double; +begin + Result := false; + FrameFinished := 0; + + if fEOF then + Exit; + + // read packets until we have a finished frame (or there are no more packets) + while (FrameFinished = 0) do + begin + errnum := av_read_frame(fFormatContext, AVPacket); + if (errnum < 0) then + begin + // failed to read a frame, check reason + + {$IF (LIBAVFORMAT_VERSION_MAJOR >= 52)} + pbIOCtx := fFormatContext^.pb; + {$ELSE} + pbIOCtx := @fFormatContext^.pb; + {$IFEND} + + // check for end-of-file (EOF is not an error) + if (url_feof(pbIOCtx) <> 0) then + begin + fEOF := true; + Exit; + end; + + // check for errors + if (url_ferror(pbIOCtx) <> 0) then + begin + Log.LogError('Video decoding file error', 'TVideoPlayback_FFmpeg.DecodeFrame'); + Exit; + end; + + // url_feof() does not detect an EOF for some mov-files (e.g. deluxe.mov) + // so we have to do it this way. + if ((fFormatContext^.file_size <> 0) and + (pbIOCtx^.pos >= fFormatContext^.file_size)) then + begin + fEOF := true; + Exit; + end; + + // error occured, log and exit + Log.LogError('Video decoding error', 'TVideoPlayback_FFmpeg.DecodeFrame'); + Exit; + end; + + // if we got a packet from the video stream, then decode it + if (AVPacket.stream_index = fStreamIndex) then + begin + // save pts to be stored in pFrame in first call of PtsGetBuffer() + VideoPktPts := AVPacket.pts; + fCodecContext^.opaque := @VideoPktPts; + + // decode packet + avcodec_decode_video(fCodecContext, fAVFrame, + frameFinished, AVPacket.data, AVPacket.size); + + // reset opaque data + fCodecContext^.opaque := nil; + + // update pts + if (AVPacket.dts <> AV_NOPTS_VALUE) then + begin + pts := AVPacket.dts; + end + else if ((fAVFrame^.opaque <> nil) and + (Pint64(fAVFrame^.opaque)^ <> AV_NOPTS_VALUE)) then + begin + pts := Pint64(fAVFrame^.opaque)^; + end + else + begin + pts := 0; + end; + + if fStream^.start_time <> AV_NOPTS_VALUE then + pts := pts - fStream^.start_time; + + pts := pts * av_q2d(fStream^.time_base); + + // synchronize time on each complete frame + if (frameFinished <> 0) then + SynchronizeTime(fAVFrame, pts); + end; + + // free the packet from av_read_frame + av_free_packet( @AVPacket ); + end; + + Result := true; +end; + +function TVideoDecodeStream_FFmpeg.GetFrame(Time: Extended): PByteArray; +var + errnum: Integer; + CurrentTime: Extended; + TimeDiff: Extended; + DropFrameCount: Integer; + i: Integer; + Success: boolean; +const + SKIP_FRAME_DIFF = 0.010; // start skipping if we are >= 10ms too late +begin + Result := nil; + + if not fOpened then + Exit; + + {* + * Synchronization - begin + *} + + // requested stream position (relative to the last loop's start) + if (fLoop) then + CurrentTime := Time - fLoopTime + else + CurrentTime := Time; + + // check if current texture still contains the active frame + if (fFrameTexValid) then + begin + // time since the last frame was returned + TimeDiff := CurrentTime - fFrameTime; + + {$IFDEF DebugDisplay} + DebugWriteln('Time: '+inttostr(floor(Time*1000)) + sLineBreak + + 'VideoTime: '+inttostr(floor(fFrameTime*1000)) + sLineBreak + + 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + + 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); + {$endif} + + // check if time has reached the next frame + if (TimeDiff < fFrameDuration) then + begin + {$ifdef DebugFrames} + // frame delay debug display + GoldenRec.Spawn(200,15,1,16,0,-1,ColoredStar,$00ff00); + {$endif} + + {$IFDEF DebugDisplay} + DebugWriteln('not getting new frame' + sLineBreak + + 'Time: '+inttostr(floor(Time*1000)) + sLineBreak + + 'VideoTime: '+inttostr(floor(fFrameTime*1000)) + sLineBreak + + 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + + 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); + {$endif} + + // we do not need a new frame now + Exit; + end; + end; + + // fetch new frame (updates fFrameTime) + Success := DecodeFrame(); + TimeDiff := CurrentTime - fFrameTime; + + // check if we have to skip frames + // Either if we are one frame behind or if the skip threshold has been reached. + // Do not skip if the difference is less than fFrameDuration as there is no next frame. + // Note: We assume that fFrameDuration is the length of one frame. + if (TimeDiff >= Max(fFrameDuration, SKIP_FRAME_DIFF)) then + begin + {$IFDEF DebugFrames} + //frame drop debug display + GoldenRec.Spawn(200,55,1,16,0,-1,ColoredStar,$ff0000); + {$ENDIF} + {$IFDEF DebugDisplay} + DebugWriteln('skipping frames' + sLineBreak + + 'TimeBase: '+inttostr(floor(fFrameDuration*1000)) + sLineBreak + + 'TimeDiff: '+inttostr(floor(TimeDifference*1000))); + {$endif} + + // update video-time + DropFrameCount := Trunc(TimeDiff / fFrameDuration); + fFrameTime := fFrameTime + DropFrameCount*fFrameDuration; + + // skip frames + for i := 1 to DropFrameCount do + Success := DecodeFrame(); + end; + + // check if we got an EOF or error + if (not Success) then + begin + if fLoop then + begin + // we have to loop, so rewind + SetPosition(0); + // record the start-time of the current loop, so we can + // determine the position in the stream (fFrameTime-fLoopTime) later. + fLoopTime := Time; + end; + Exit; + end; + + {* + * Synchronization - end + *} + + // TODO: support for pan&scan + //if (fAVFrame.pan_scan <> nil) then + //begin + // Writeln(Format('PanScan: %d/%d', [fAVFrame.pan_scan.width, fAVFrame.pan_scan.height])); + //end; + + // otherwise we convert the pixeldata from YUV to RGB + {$IFDEF UseSWScale} + errnum := sws_scale(fSwScaleContext, @fAVFrame.data, @fAVFrame.linesize, + 0, fCodecContext^.Height, + @fAVFrameRGB.data, @fAVFrameRGB.linesize); + {$ELSE} + // img_convert from lib/ffmpeg/avcodec.pas is actually deprecated. + // If ./configure does not find SWScale then this gives the error + // that the identifier img_convert is not known or similar. + // I think this should be removed, but am not sure whether there should + // be some other replacement or a warning, Therefore, I leave it for now. + // April 2009, mischi + errnum := img_convert(PAVPicture(fAVFrameRGB), PIXEL_FMT_FFMPEG, + PAVPicture(fAVFrame), fCodecContext^.pix_fmt, + fCodecContext^.width, fCodecContext^.height); + {$ENDIF} + + if (errnum < 0) then + begin + Log.LogError('Image conversion failed', 'TVideoPlayback_ffmpeg.GetFrame'); + Exit; + end; + + if (not fFrameTexValid) then + fFrameTexValid := true; + + Result := PByteArray(fAVFrameRGB^.data[0]); +end; + +procedure TVideoDecodeStream_FFmpeg.SetLoop(Enable: boolean); +begin + fLoop := Enable; + fLoopTime := 0; +end; + +function TVideoDecodeStream_FFmpeg.GetLoop(): boolean; +begin + Result := fLoop; +end; + +{** + * Sets the stream's position. + * The stream is set to the first keyframe with timestamp <= Time. + * Note that fFrameTime is set to Time no matter if the actual position seeked to is + * at Time or the time of a preceding keyframe. fFrameTime will be updated to the + * actual frame time when GetFrame() is called the next time. + * @param Time new position in seconds + *} +procedure TVideoDecodeStream_FFmpeg.SetPosition(Time: real); +var + SeekFlags: integer; +begin + if not fOpened then + Exit; + + if (Time < 0) then + Time := 0; + + // TODO: handle fLoop-times + //Time := Time mod VideoDuration; + + // Do not use the AVSEEK_FLAG_ANY here. It will seek to any frame, even + // non keyframes (P-/B-frames). It will produce corrupted video frames as + // FFmpeg does not use the information of the preceding I-frame. + // The picture might be gray or green until the next keyframe occurs. + // Instead seek the first keyframe smaller than the requested time + // (AVSEEK_FLAG_BACKWARD). As this can be some seconds earlier than the + // requested time, let the sync in GetFrame() do its job. + SeekFlags := AVSEEK_FLAG_BACKWARD; + + fFrameTime := Time; + fEOF := false; + fFrameTexValid := false; + + if (av_seek_frame(fFormatContext, + fStreamIndex, + Round(Time / av_q2d(fStream^.time_base)), + SeekFlags) < 0) then + begin + Log.LogError('av_seek_frame() failed', 'TVideoPlayback_ffmpeg.SetPosition'); + Exit; + end; + + avcodec_flush_buffers(fCodecContext); +end; + +function TVideoDecodeStream_FFmpeg.GetPosition: real; +begin + Result := fFrameTime; +end; + +function TVideoDecodeStream_FFmpeg.GetFrameWidth(): integer; +begin + Result := fCodecContext^.width; +end; + +function TVideoDecodeStream_FFmpeg.GetFrameHeight(): integer; +begin + Result := fCodecContext^.height; +end; + +function TVideoDecodeStream_FFmpeg.GetFrameAspect(): real; +begin + Result := fAspect; +end; + +initialization + MediaManager.Add(TVideoDecoder_FFmpeg.Create); + +end. |