diff options
Diffstat (limited to '')
-rw-r--r-- | src/lib/ffmpeg-2.6/avcodec.pas | 129 | ||||
-rw-r--r-- | src/lib/ffmpeg-2.6/avformat.pas | 91 | ||||
-rw-r--r-- | src/lib/ffmpeg-2.6/avio.pas | 7 | ||||
-rw-r--r-- | src/lib/ffmpeg-2.6/ff_api-defines.inc | 6 | ||||
-rw-r--r-- | src/lib/ffmpeg-2.6/libavutil/frame.pas | 12 | ||||
-rw-r--r-- | src/lib/ffmpeg-2.6/libavutil/pixfmt.pas | 56 |
6 files changed, 205 insertions, 96 deletions
diff --git a/src/lib/ffmpeg-2.6/avcodec.pas b/src/lib/ffmpeg-2.6/avcodec.pas index a1e4aa02..18bb091a 100644 --- a/src/lib/ffmpeg-2.6/avcodec.pas +++ b/src/lib/ffmpeg-2.6/avcodec.pas @@ -338,6 +338,7 @@ type AV_CODEC_ID_SGIRLE_DEPRECATED, AV_CODEC_ID_MVC1_DEPRECATED, AV_CODEC_ID_MVC2_DEPRECATED, + AV_CODEC_ID_HQX, (** see below. they need to be hard coded. AV_CODEC_ID_BRENDER_PIX= MKBETAG('B','P','I','X'), @@ -440,7 +441,9 @@ type AV_CODEC_ID_ADPCM_VIMA_DEPRECATED, (** see below. they need to be hard coded. AV_CODEC_ID_ADPCM_VIMA = MKBETAG('V','I','M','A'), +{$IFDEF FF_API_VIMA_DECODER} AV_CODEC_ID_VIMA = MKBETAG('V','I','M','A'), +{$IFEND} AV_CODEC_ID_ADPCM_AFC = MKBETAG('A','F','C',' '), AV_CODEC_ID_ADPCM_IMA_OKI = MKBETAG('O','K','I',' '), AV_CODEC_ID_ADPCM_DTK = MKBETAG('D','T','K',' '), @@ -531,6 +534,7 @@ type AV_CODEC_ID_TAK_DEPRECATED, AV_CODEC_ID_PAF_AUDIO_DEPRECATED, AV_CODEC_ID_ON2AVC, + AV_CODEC_ID_DSS_SP, (** see below. they need to be hard coded. AV_CODEC_ID_FFWAVESYNTH = MKBETAG('F','F','W','S'), AV_CODEC_ID_SONIC = MKBETAG('S','O','N','C'), @@ -1451,6 +1455,12 @@ type AV_PKT_DATA_STEREO3D, (** + * This side data should be associated with an audio stream and corresponds + * to enum AVAudioServiceType. + *) + AV_PKT_DATA_AUDIO_SERVICE_TYPE, + + (** * Recommmends skipping the specified number of samples * @code * u32le number of samples to skip from start of this packet @@ -2171,13 +2181,13 @@ type *) codec_tag: cuint; +{$IFDEF FF_API_STREAM_CODEC_TAG} (** - * fourcc from the AVI stream header (LSB first, so "ABCD" -> ('D'<<24) + ('C'<<16) + ('B'<<8) + 'A'). - * This is used to work around some encoder bugs. - * - encoding: unused - * - decoding: Set by user, will be converted to uppercase by libavcodec during init. + * @deprecated this field is unused *) + {attribute_deprecated} stream_codec_tag: cuint; +{$IFEND} priv_data: pointer; @@ -3662,6 +3672,13 @@ type framerate: TAVRational; (** + * Nominal unaccelerated pixel format, see AV_PIX_FMT_xxx. + * - encoding: unused. + * - decoding: Set by libavcodec before calling get_format() + *) + sw_pix_fmt: TAVPixelFormat; + + (** * Timebase in which pkt_dts/pts and AVPacket.dts/pts are. * Code outside libavcodec should access this field using: * av_codec_{get,set}_pkt_timebase(avctx) @@ -3829,7 +3846,7 @@ type (** * data+linesize for the bitmap of this subtitle. - * can be set for text/ass as well once they where rendered + * can be set for text/ass as well once they are rendered *) pict: TAVPicture; type_: TAVSubtitleType; @@ -4107,6 +4124,12 @@ const *) AV_HWACCEL_FLAG_IGNORE_LEVEL = (1 << 0); + (** + * Hardware acceleration can output YUV pixel formats with a different chroma + * sampling than 4:2:0 and/or other than 8 bits per component. + *) + AV_HWACCEL_FLAG_ALLOW_HIGH_DEPTH = (1 << 1); + FF_SUB_CHARENC_MODE_DO_NOTHING = -1; ///< do nothing (demuxer outputs a stream supposed to be already in UTF-8, or the codec is bitmap for instance) FF_SUB_CHARENC_MODE_AUTOMATIC = 0; ///< libavcodec will select the mode itself FF_SUB_CHARENC_MODE_PRE_DECODER = 1; ///< the AVPacket data needs to be recoded to UTF-8 before being fed to the decoder, requires iconv @@ -4315,6 +4338,9 @@ procedure avcodec_free_frame(frame: PPAVFrame); * * @warning This function is not thread safe! * + * @note Always call this function before using decoding routines (such as + * @ref avcodec_decode_video2()). + * * @code * avcodec_register_all(); * av_dict_set(&opts, "b", "2.5M", 0); @@ -4756,18 +4782,27 @@ function avcodec_decode_audio3(avctx: PAVCodecContext; samples: PSmallint; * Decode the audio frame of size avpkt->size from avpkt->data into frame. * * Some decoders may support multiple frames in a single AVPacket. Such - * decoders would then just decode the first frame. In this case, - * avcodec_decode_audio4 has to be called again with an AVPacket containing - * the remaining data in order to decode the second frame, etc... - * Even if no frames are returned, the packet needs to be fed to the decoder - * with remaining data until it is completely consumed or an error occurs. + * decoders would then just decode the first frame and the return value would be + * less than the packet size. In this case, avcodec_decode_audio4 has to be + * called again with an AVPacket containing the remaining data in order to + * decode the second frame, etc... Even if no frames are returned, the packet + * needs to be fed to the decoder with remaining data until it is completely + * consumed or an error occurs. + * + * Some decoders (those marked with CODEC_CAP_DELAY) have a delay between input + * and output. This means that for some packets they will not immediately + * produce decoded output and need to be flushed at the end of decoding to get + * all the decoded data. Flushing is done by calling this function with packets + * with avpkt->data set to NULL and avpkt->size set to 0 until it stops + * returning samples. It is safe to flush even those decoders that are not + * marked with CODEC_CAP_DELAY, then no samples will be returned. * * @warning The input buffer, avpkt->data must be FF_INPUT_BUFFER_PADDING_SIZE * larger than the actual read bytes because some optimized bitstream * readers read 32 or 64 bits at once and could read over the end. * - * @note You might have to align the input buffer. The alignment requirements - * depend on the CPU and the decoder. + * @note The AVCodecContext MUST have been opened with @ref avcodec_open2() + * before packets may be fed to the decoder. * * @param avctx the codec context * @param[out] frame The AVFrame in which to store decoded audio samples. @@ -4780,10 +4815,13 @@ function avcodec_decode_audio3(avctx: PAVCodecContext; samples: PSmallint; * to the frame if av_frame_is_writable() returns 1. * When AVCodecContext.refcounted_frames is set to 0, the returned * reference belongs to the decoder and is valid only until the - * next call to this function or until closing the decoder. - * The caller may not write to it. + * next call to this function or until closing or flushing the + * decoder. The caller may not write to it. * @param[out] got_frame_ptr Zero if no frame could be decoded, otherwise it is - * non-zero. + * non-zero. Note that this field being set to zero + * does not mean that an error has occurred. For + * decoders with CODEC_CAP_DELAY set, no given decode + * call is guaranteed to produce a frame. * @param[in] avpkt The input AVPacket containing the input buffer. * At least avpkt->data and avpkt->size should be set. Some * decoders might also require additional fields to be set. @@ -4807,17 +4845,13 @@ function avcodec_decode_audio4(avctx: PAVCodecContext; frame: PAVFrame; * @warning The end of the input buffer buf should be set to 0 to ensure that * no overreading happens for damaged MPEG streams. * - * @note You might have to align the input buffer avpkt->data. - * The alignment requirements depend on the CPU: on some CPUs it isn't - * necessary at all, on others it won't work at all if not aligned and on others - * it will work but it will have an impact on performance. - * - * In practice, avpkt->data should have 4 byte alignment at minimum. - * * @note Codecs which have the CODEC_CAP_DELAY capability set have a delay * between input and output, these need to be fed with avpkt->data=NULL, * avpkt->size=0 at the end to return the remaining frames. * + * @note The AVCodecContext MUST have been opened with @ref avcodec_open2() + * before packets may be fed to the decoder. + * * @param avctx the codec context * @param[out] picture The AVFrame in which the decoded video frame will be stored. * Use av_frame_alloc() to get an AVFrame. The codec will @@ -4830,10 +4864,10 @@ function avcodec_decode_audio4(avctx: PAVCodecContext; frame: PAVFrame; * to the frame if av_frame_is_writable() returns 1. * When AVCodecContext.refcounted_frames is set to 0, the returned * reference belongs to the decoder and is valid only until the - * next call to this function or until closing the decoder. The - * caller may not write to it. + * next call to this function or until closing or flushing the + * decoder. The caller may not write to it. * - * @param[in] avpkt The input AVpacket containing the input buffer. + * @param[in] avpkt The input AVPacket containing the input buffer. * You can create such packet with av_init_packet() and by then setting * data and size, some decoders might in addition need other fields like * flags&AV_PKT_FLAG_KEY. All decoders are designed to use the least @@ -4847,7 +4881,8 @@ function avcodec_decode_video2(avctx: PAVCodecContext; picture: PAVFrame; avpkt: {const} PAVPacket): cint; cdecl; external av__codec; -(* Decode a subtitle message. +(* + * Decode a subtitle message. * Return a negative value on error, otherwise return the number of bytes used. * If no subtitle could be decompressed, got_sub_ptr is zero. * Otherwise, the subtitle is stored in *sub. @@ -4856,6 +4891,17 @@ function avcodec_decode_video2(avctx: PAVCodecContext; picture: PAVFrame; * and reusing a get_buffer written for video codecs would probably perform badly * due to a potentially very different allocation pattern. * + * Some decoders (those marked with CODEC_CAP_DELAY) have a delay between input + * and output. This means that for some packets they will not immediately + * produce decoded output and need to be flushed at the end of decoding to get + * all the decoded data. Flushing is done by calling this function with packets + * with avpkt->data set to NULL and avpkt->size set to 0 until it stops + * returning subtitles. It is safe to flush even those decoders that are not + * marked with CODEC_CAP_DELAY, then no subtitles will be returned. + * + * @note The AVCodecContext MUST have been opened with @ref avcodec_open2() + * before packets may be fed to the decoder. + * * @param avctx the codec context * @param[out] sub The Preallocated AVSubtitle in which the decoded subtitle will be stored, * must be freed with avsubtitle_free if *got_sub_ptr is set. @@ -5032,12 +5078,43 @@ type *) picture_structure: TAVPictureStructure; + (** + * Picture number incremented in presentation or output order. + * This field may be reinitialized at the first picture of a new sequence. + * + * For example, this corresponds to H.264 PicOrderCnt. + *) + output_picture_number: cint; + + (** + * Dimensions of the decoded video intended for presentation. + *) + width: cint; + height: cint; + + (** + * Dimensions of the coded video. + *) + coded_width: cint; + coded_height: cint; + + (** + * The format of the coded data, corresponds to enum AVPixelFormat for video + * and for enum AVSampleFormat for audio. + * + * Note that a decoder can have considerable freedom in how exactly it + * decodes the data, so the format reported here might be different from the + * one returned by a decoder. + *) + format: cint; end; {AVCodecParserContext} TAVCodecParser = record codec_ids: array [0..4] of cint; (* several codec IDs are permitted *) priv_data_size: cint; parser_init: function(s: PAVCodecParserContext): cint; cdecl; + (* This callback never returns an error, a negative value means that + * the frame start was in a previous packet. *) parser_parse: function(s: PAVCodecParserContext; avctx: PAVCodecContext; poutbuf: {const} PPointer; poutbuf_size: PCint; diff --git a/src/lib/ffmpeg-2.6/avformat.pas b/src/lib/ffmpeg-2.6/avformat.pas index 660cf3ce..fdf97dca 100644 --- a/src/lib/ffmpeg-2.6/avformat.pas +++ b/src/lib/ffmpeg-2.6/avformat.pas @@ -264,6 +264,10 @@ const * be set to the timebase that the caller desires to use for this stream (note * that the timebase actually used by the muxer can be different, as will be * described later). + * - It is advised to manually initialize only the relevant fields in + * AVCodecContext, rather than using @ref avcodec_copy_context() during + * remuxing: there is no guarantee that the codec context values remain valid + * for both input and output format contexts. * - The caller may fill in additional information, such as @ref * AVFormatContext.metadata "global" or @ref AVStream.metadata "per-stream" * metadata, @ref AVFormatContext.chapters "chapters", @ref @@ -755,6 +759,8 @@ type * @see avdevice_capabilities_free() for more details. *) free_device_capabilities: function(s: PAVFormatContext; caps: PAVDeviceCapabilitiesQuery): cint; cdecl; + + data_codec: TAVCodecID; (**< default data codec *) end; (** * @} @@ -1734,54 +1740,6 @@ type *) format_whitelist: PAnsiChar; - (***************************************************************** - * All fields below this line are not part of the public API. They - * may not be used outside of libavformat and can be changed and - * removed at will. - * New public fields should be added right above. - ***************************************************************** - *) - - (** - * This buffer is only needed when packets were already buffered but - * not decoded, for example to get the codec parameters in MPEG - * streams. - *) - packet_buffer: PAVPacketList; - packet_buffer_end_: PAVPacketList; - - (* av_seek_frame() support *) - data_offset: cint64; (**< offset of the first packet *) - - (** - * Raw packets from the demuxer, prior to parsing and decoding. - * This buffer is used for buffering packets until the codec can - * be identified, as parsing cannot be done without knowing the - * codec. - *) - raw_packet_buffer_: PAVPacketList; - raw_packet_buffer_end_: PAVPacketList; - (** - * Packets split by the parser get queued here. - *) - parse_queue: PAVPacketList; - parse_queue_end: PAVPacketList; - (** - * Remaining size available for raw_packet_buffer, in bytes. - *) - raw_packet_buffer_remaining_size: cint; - - (** - * Offset to remap timestamps to be non-negative. - * Expressed in timebase units. - * @see AVStream.mux_ts_offset - *) - offset: cint64; - - (** - * Timebase for the timestamp offset. - *) - offset_timebase: TAVRational; (** * An opaque field for libavformat internal usage. @@ -1822,6 +1780,14 @@ type subtitle_codec: PAVCodec; (** + * Forced data codec. + * This allows forcing a specific decoder, even when there are multiple with + * the same codec_id. + * Demuxing: Set by user via av_format_set_data_codec (NO direct access). + *) + data_codec: PAVCodec; + + (** * Number of bytes to be written as padding in a metadata header. * Demuxing: Unused. * Muxing: Set by user via av_format_set_metadata_header_padding. @@ -1872,6 +1838,12 @@ type * - demuxing: Set by user. *) dump_separator: Pcuint8; + + (** + * Forced Data codec_id. + * Demuxing: Set by user. + *) + data_codec_id: TAVCodecID; end; (** TAVFormatContext **) function av_format_get_probe_score(s: {const} PAVFormatContext): cint; @@ -1888,6 +1860,10 @@ function av_format_get_subtitle_codec(s: {const} PAVFormatContext): PAVCodec; cdecl; external av__format; procedure av_format_set_subtitle_codec(s: PAVFormatContext; c: PAVCodec); cdecl; external av__format; +function av_format_get_data_codec(s: {const} PAVFormatContext): PAVCodec; + cdecl; external av__format; +procedure av_format_set_data_codec(s: PAVFormatContext; c: PAVCodec); + cdecl; external av__format; function av_format_get_metadata_header_padding(s: {const} PAVFormatContext): cint; cdecl; external av__format; procedure av_format_set_metadata_header_padding(s: PAVFormatContext; c: cint); @@ -2340,6 +2316,25 @@ function avformat_seek_file(s: PAVFormatContext; stream_index: cint; min_ts, ts, cdecl; external av__format; (** + * Discard all internally buffered data. This can be useful when dealing with + * discontinuities in the byte stream. Generally works only with formats that + * can resync. This includes headerless formats like MPEG-TS/TS but should also + * work with NUT, Ogg and in a limited way AVI for example. + * + * The set of streams, the detected duration, stream parameters and codecs do + * not change when calling this function. If you want a complete reset, it's + * better to open a new AVFormatContext. + * + * This does not flush the AVIOContext (s->pb). If necessary, call + * avio_flush(s->pb) before calling this function. + * + * @param s media file handle + * @return >=0 on success, error code otherwise + *) +function avformat_flush(s: PAVFormatContext): cint; + cdecl; external av__format; + +(** * Start playing a network-based stream (e.g. RTSP stream) at the * current position. *) diff --git a/src/lib/ffmpeg-2.6/avio.pas b/src/lib/ffmpeg-2.6/avio.pas index c342f8de..db3a84e6 100644 --- a/src/lib/ffmpeg-2.6/avio.pas +++ b/src/lib/ffmpeg-2.6/avio.pas @@ -276,6 +276,13 @@ function avio_put_str(s: PAVIOContext; str: {const} PAnsiChar): cint; function avio_put_str16le(s: PAVIOContext; str: {const} PAnsiChar): cint; cdecl; external av__format; +(** + * Convert an UTF-8 string to UTF-16BE and write it. + * @return number of bytes written. + *) +function avio_put_str16be(s: PAVIOContext; str: {const} PAnsiChar): cint; + cdecl; external av__format; + const (** * Passing this as the "whence" parameter to a seek function causes it to diff --git a/src/lib/ffmpeg-2.6/ff_api-defines.inc b/src/lib/ffmpeg-2.6/ff_api-defines.inc index a964caaf..5c9b4bb2 100644 --- a/src/lib/ffmpeg-2.6/ff_api-defines.inc +++ b/src/lib/ffmpeg-2.6/ff_api-defines.inc @@ -8,6 +8,9 @@ *) (** avcodec defines *) +{$ifndef FF_API_VIMA_DECODER} +{$define FF_API_VIMA_DECODER := (LIBAVCODEC_VERSION_MAJOR < 57)} +{$endif} {$ifndef FF_API_REQUEST_CHANNELS} {$define FF_API_REQUEST_CHANNELS := (LIBAVCODEC_VERSION_MAJOR < 57)} {$endif} @@ -144,6 +147,9 @@ {$ifndef FF_API_MPV_OPT} {$define FF_API_MPV_OPT := (LIBAVCODEC_VERSION_MAJOR < 59)} {$endif} +{$ifndef FF_API_STREAM_CODEC_TAG} +{$define FF_API_STREAM_CODEC_TAG := (LIBAVCODEC_VERSION_MAJOR < 59)} +{$endif} (* avutil defines *) {$ifndef FF_API_GET_BITS_PER_SAMPLE_FMT} diff --git a/src/lib/ffmpeg-2.6/libavutil/frame.pas b/src/lib/ffmpeg-2.6/libavutil/frame.pas index 87eb21ff..f47f2397 100644 --- a/src/lib/ffmpeg-2.6/libavutil/frame.pas +++ b/src/lib/ffmpeg-2.6/libavutil/frame.pas @@ -157,7 +157,13 @@ type * u8 reason for end skip (0=padding silence, 1=convergence) * @endcode *) - AV_FRAME_DATA_SKIP_SAMPLES + AV_FRAME_DATA_SKIP_SAMPLES, + + (** + * This side data must be associated with an audio frame and corresponds to + * enum AVAudioServiceType defined in avcodec.h. + *) + AV_FRAME_DATA_AUDIO_SERVICE_TYPE ); TAVActiveFormatDescription = ( @@ -452,7 +458,9 @@ type (** * AVBuffer references backing the data for this frame. If all elements of - * this array are NULL, then this frame is not reference counted. + * this array are NULL, then this frame is not reference counted. This array + * must be filled contiguously -- if buf[i] is non-NULL then buf[j] must + * also be non-NULL for all j < i. * * There may be at most one AVBuffer per data plane, so for video this array * always contains all the references. For planar audio with more than diff --git a/src/lib/ffmpeg-2.6/libavutil/pixfmt.pas b/src/lib/ffmpeg-2.6/libavutil/pixfmt.pas index f0539c9a..7709d7f7 100644 --- a/src/lib/ffmpeg-2.6/libavutil/pixfmt.pas +++ b/src/lib/ffmpeg-2.6/libavutil/pixfmt.pas @@ -44,11 +44,11 @@ type * big-endian CPUs. * * @par - * When the pixel format is palettized RGB (AV_PIX_FMT_PAL8), the palettized + * When the pixel format is palettized RGB32 (AV_PIX_FMT_PAL8), the palettized * image data is stored in AVFrame.data[0]. The palette is transported in * AVFrame.data[1], is 1024 bytes long (256 4-byte entries) and is * formatted the same as in AV_PIX_FMT_RGB32 described above (i.e., it is - * also endian-specific). Note also that the individual RGB palette + * also endian-specific). Note also that the individual RGB32 palette * components stored in AVFrame.data[1] should be in the range 0..255. * This is important as many custom PAL8 video codecs that were designed * to run on the IBM VGA graphics adapter use 6-bit palette components. @@ -120,13 +120,13 @@ type AV_PIX_FMT_RGB565BE, ///< packed RGB 5:6:5, 16bpp, (msb) 5R 6G 5B(lsb), big-endian AV_PIX_FMT_RGB565LE, ///< packed RGB 5:6:5, 16bpp, (msb) 5R 6G 5B(lsb), little-endian - AV_PIX_FMT_RGB555BE, ///< packed RGB 5:5:5, 16bpp, (msb)1A 5R 5G 5B(lsb), big-endian, most significant bit to 0 - AV_PIX_FMT_RGB555LE, ///< packed RGB 5:5:5, 16bpp, (msb)1A 5R 5G 5B(lsb), little-endian, most significant bit to 0 + AV_PIX_FMT_RGB555BE, ///< packed RGB 5:5:5, 16bpp, (msb)1X 5R 5G 5B(lsb), big-endian, most significant bit to 0 + AV_PIX_FMT_RGB555LE, ///< packed RGB 5:5:5, 16bpp, (msb)1X 5R 5G 5B(lsb), little-endian, most significant bit to 0 AV_PIX_FMT_BGR565BE, ///< packed BGR 5:6:5, 16bpp, (msb) 5B 6G 5R(lsb), big-endian AV_PIX_FMT_BGR565LE, ///< packed BGR 5:6:5, 16bpp, (msb) 5B 6G 5R(lsb), little-endian - AV_PIX_FMT_BGR555BE, ///< packed BGR 5:5:5, 16bpp, (msb)1A 5B 5G 5R(lsb), big-endian, most significant bit to 1 - AV_PIX_FMT_BGR555LE, ///< packed BGR 5:5:5, 16bpp, (msb)1A 5B 5G 5R(lsb), little-endian, most significant bit to 1 + AV_PIX_FMT_BGR555BE, ///< packed BGR 5:5:5, 16bpp, (msb)1X 5B 5G 5R(lsb), big-endian, most significant bit to 1 + AV_PIX_FMT_BGR555LE, ///< packed BGR 5:5:5, 16bpp, (msb)1X 5B 5G 5R(lsb), little-endian, most significant bit to 1 AV_PIX_FMT_VAAPI_MOCO, ///< HW acceleration through VA API at motion compensation entry-point, Picture.data[3] contains a vaapi_render_state struct which contains macroblocks as well as various fields extracted from headers AV_PIX_FMT_VAAPI_IDCT, ///< HW acceleration through VA API at IDCT entry-point, Picture.data[3] contains a vaapi_render_state struct which contains fields extracted from headers @@ -143,10 +143,10 @@ type {$ENDIF} AV_PIX_FMT_DXVA2_VLD, ///< HW decoding through DXVA2, Picture.data[3] contains a LPDIRECT3DSURFACE9 pointer - AV_PIX_FMT_RGB444LE, ///< packed RGB 4:4:4, 16bpp, (msb)4A 4R 4G 4B(lsb), little-endian, most significant bits to 0 - AV_PIX_FMT_RGB444BE, ///< packed RGB 4:4:4, 16bpp, (msb)4A 4R 4G 4B(lsb), big-endian, most significant bits to 0 - AV_PIX_FMT_BGR444LE, ///< packed BGR 4:4:4, 16bpp, (msb)4A 4B 4G 4R(lsb), little-endian, most significant bits to 1 - AV_PIX_FMT_BGR444BE, ///< packed BGR 4:4:4, 16bpp, (msb)4A 4B 4G 4R(lsb), big-endian, most significant bits to 1 + AV_PIX_FMT_RGB444LE, ///< packed RGB 4:4:4, 16bpp, (msb)4X 4R 4G 4B(lsb), little-endian, most significant bits to 0 + AV_PIX_FMT_RGB444BE, ///< packed RGB 4:4:4, 16bpp, (msb)4X 4R 4G 4B(lsb), big-endian, most significant bits to 0 + AV_PIX_FMT_BGR444LE, ///< packed BGR 4:4:4, 16bpp, (msb)4X 4B 4G 4R(lsb), little-endian, most significant bits to 1 + AV_PIX_FMT_BGR444BE, ///< packed BGR 4:4:4, 16bpp, (msb)4X 4B 4G 4R(lsb), big-endian, most significant bits to 1 AV_PIX_FMT_YA8, ///< 8bit gray, 8bit alpha (* see const declaration way down AV_PIX_FMT_Y400A = AV_PIX_FMT_YA8, ///< alias for AV_PIX_FMT_YA8 @@ -241,6 +241,19 @@ type AV_PIX_FMT_YA16BE, ///< 16bit gray, 16bit alpha (big-endian) AV_PIX_FMT_YA16LE, ///< 16bit gray, 16bit alpha (little-endian) + (** + * duplicated pixel formats for compatibility with libav. + * FFmpeg supports these formats since May 3 2013 (commit e6d4e687558d08187e7a415a7725e4b1a416f782) + * Libav added them Jan 14 2015 with incompatible values (commit 0e6c7dfa650e8b0497bfa7a06394b7a462ddc33a) + *) + AV_PIX_FMT_GBRAP_LIBAV, ///< planar GBRA 4:4:4:4 32bpp + AV_PIX_FMT_GBRAP16BE_LIBAV, ///< planar GBRA 4:4:4:4 64bpp, big-endian + AV_PIX_FMT_GBRAP16LE_LIBAV, ///< planar GBRA 4:4:4:4 64bpp, little-endian + (** + * HW acceleration through QSV, data[3] contains a pointer to the + * mfxFrameSurface1 structure. + *) + AV_PIX_FMT_QSV, {$IFNDEF AV_PIX_FMT_ABI_GIT_MASTER} AV_PIX_FMT_RGBA64BE = $123, ///< packed RGBA 16:16:16:16, 64bpp, 16R, 16G, 16B, 16A, the 2-byte value for each R/G/B/A component is stored as big-endian @@ -248,10 +261,10 @@ type AV_PIX_FMT_BGRA64BE, ///< packed RGBA 16:16:16:16, 64bpp, 16B, 16G, 16R, 16A, the 2-byte value for each R/G/B/A component is stored as big-endian AV_PIX_FMT_BGRA64LE, ///< packed RGBA 16:16:16:16, 64bpp, 16B, 16G, 16R, 16A, the 2-byte value for each R/G/B/A component is stored as little-endian {$ENDIF} - AV_PIX_FMT_0RGB = $123 + 4, ///< packed RGB 8:8:8, 32bpp, 0RGB0RGB... - AV_PIX_FMT_RGB0, ///< packed RGB 8:8:8, 32bpp, RGB0RGB0... - AV_PIX_FMT_0BGR, ///< packed BGR 8:8:8, 32bpp, 0BGR0BGR... - AV_PIX_FMT_BGR0, ///< packed BGR 8:8:8, 32bpp, BGR0BGR0... + AV_PIX_FMT_0RGB = $123 + 4,///< packed RGB 8:8:8, 32bpp, XRGBXRGB... X=unused/undefined + AV_PIX_FMT_RGB0, ///< packed RGB 8:8:8, 32bpp, RGBXRGBX... X=unused/undefined + AV_PIX_FMT_0BGR, ///< packed BGR 8:8:8, 32bpp, XBGRXBGR... X=unused/undefined + AV_PIX_FMT_BGR0, ///< packed BGR 8:8:8, 32bpp, BGRXBGRX... X=unused/undefined AV_PIX_FMT_YUVA444P, ///< planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples) AV_PIX_FMT_YUVA422P, ///< planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples) @@ -296,12 +309,15 @@ type {$IFDEF AV_HAVE_INCOMPATIBLE_LIBAV_ABI} const - AV_PIX_FMT_YUVA422P = AV_PIX_FMT_YUVA422P_LIBAV; - AV_PIX_FMT_YUVA444P = AV_PIX_FMT_YUVA444P_LIBAV; - AV_PIX_FMT_RGBA64BE = AV_PIX_FMT_RGBA64BE_LIBAV; - AV_PIX_FMT_RGBA64LE = AV_PIX_FMT_RGBA64LE_LIBAV; - AV_PIX_FMT_BGRA64BE = AV_PIX_FMT_BGRA64BE_LIBAV; - AV_PIX_FMT_BGRA64LE = AV_PIX_FMT_BGRA64LE_LIBAV; + AV_PIX_FMT_YUVA422P = AV_PIX_FMT_YUVA422P_LIBAV; + AV_PIX_FMT_YUVA444P = AV_PIX_FMT_YUVA444P_LIBAV; + AV_PIX_FMT_RGBA64BE = AV_PIX_FMT_RGBA64BE_LIBAV; + AV_PIX_FMT_RGBA64LE = AV_PIX_FMT_RGBA64LE_LIBAV; + AV_PIX_FMT_BGRA64BE = AV_PIX_FMT_BGRA64BE_LIBAV; + AV_PIX_FMT_BGRA64LE = AV_PIX_FMT_BGRA64LE_LIBAV; + AV_PIX_FMT_GBRAP = AV_PIX_FMT_GBRAP_LIBAV; + AV_PIX_FMT_GBRAP16BE = AV_PIX_FMT_GBRAP16BE_LIBAV; + AV_PIX_FMT_GBRAP16LE = AV_PIX_FMT_GBRAP16LE_LIBAV; {$ENDIF} const |