* master: mmsh: fixed printf injection bug in mmsh request ac3enc: use correct alignment and length in channel coupling dsp functions. ffmpeg: don't abuse a global for passing framerate from input to output ffmpeg: don't abuse a global for passing channels from input to output ffmpeg: don't abuse a global for passing samplerate from input to output Make buffer size check consistent and avoid a possible overflow. Fix spelling. Full support for sending H.264 in RTP ARM: update ff_h264_idct8_add4_neon for 4:4:4 changes swscale: use SwsContext for av_log when available Support reading chan atoms with empty channel descriptions. Reindent after last commit. Fix multi-channel AAC encoding. Fix "redundant redeclaration" warning. Fix compilation with --disable-everything --enable-encoder=ac3/ac3_fixed. vf_mp: Fix large memleak. swscale: Remove HAVE_MMX from files that are only compiled with MMX enabled. swscale: Fix compilation with --disable-mmx2. mjpegenc: Fix JFIF version swscale: remove misplaced comment. ffmpeg: fix streaming to ffserver. swscale: split out RGB48 output functions from yuv2packed[12X]_c(). build: move vpath directives to main Makefile swscale: fix JPEG-range YUV scaling artifacts. build: move ALLFFLIBS to a more logical place ARM: factor some repetitive code into macros CrystalHD: Use mp4toannexb bitstream filter. CrystalHD: Keep mp4toannexb filter around for entire decoder lifetime. Fix SVQ3 after adding 4:4:4 H.264 support H.264: fix CODEC_FLAG_GRAY 4:4:4 H.264 decoding support matroskadec: properly decode color space in an endian neutral way matroskadec: use a temporary fourcc variable matroskaenc: ensure the written colorspace don't depend on host endianness ac3enc: fix allocation of floating point samples. utils: Drop pointless '#if 1' preprocessor directive. ac3enc: remove empty ac3_float function that is never called ac3enc: split templated float vs. fixed functions into a separate file. ac3enc: dynamically allocate AC3EncodeContext fields windowed_samples and mdct ac3enc: use function pointer to choose between AC-3 and E-AC-3 header output functions. Roll back 4:4:4 H.264 for now Needs some ARM/PPC asm modifications. Fix SVQ3 after adding 4:4:4 H.264 support H.264: fix CODEC_FLAG_GRAY 4:4:4 H.264 decoding support h264_parser: Fix whitespace after previous change. h264_parser: Fix behaviour when PARSER_FLAG_COMPLETE_FRAMES is set. wav: remove an invalid free(). lavf: initialise reference_dts in av_estimate_timings_from_pts. h264: don't be so picky on decoding pps in extradata. avcodec.h: add or elaborate on some documentation comments. h264: change a few comments into error messages ac3dec: fix doxy-style for comment ("///>" should be "///<" instead). img2: add .dpx to the list of supported file extensions. ffv1: fix undefined behavior with insane widths. replace remaining usage of deprecated av_metadata_set2() by av_dict_set() matroskaenc: write colourspace element for rawvideo tracks nsv: simplify probe function nsv: return error code instead of discarding it in read_header() ARM: jrevdct_arm: simplify stack usage ARM: jrevdct_arm: use push/pop mnemonics ARM: jrevdct_arm: misc cleanup ARM: optimised mpadsp_apply_window_fixed Add some (important) changelog entries H264: Reduce pointless diffs to qatar Revert "H264: Split out hl_motion and template it, this seems a bit faster" libavfilter: implement avfilter_fill_frame_from_video_buffer_ref() avfiltergraph: make the AVFilterInOut alloc/free API public avfiltergraph: change the syntax of avfilter_graph_parse() graphparser: prefer void * over AVClass * for log contexts h264: Complexify frame num gap shortening code Update todo mpeg12: replace 2 asserts by av_assert0 cmdutils: add missing NULL check in parse_options() x11grab: remove a memory allocation and the associated memcpy. Fix --disable-everything build: fix "make install" with documentation disabled build: simplify some conditional targets resample: clarify supported resampling. lavfi: fix signature for avfilter_graph_parse() and avfilter_graph_config() avfiltergraph: use meaningful error codes Revert "ac3: there was no libav in 2010 thus this code cannot be from libav." Fix -t option for formats which holds dts and no pts dnxhd: Renama tables Extract rotation in MOV metadata bitstream: Properly promote av_reverse values before shifting. pixfmt: Replace 9/10bit deprecation by a technical explanation. libavutil/swscale: YUV444P10/YUV444P9 support. H.264: Fix high bit depth explicit biweight h264: Fix 10-bit H.264 x86 chroma v loopfilter asm. Replace DEBUG_SEEK/DEBUG_SI + av_log combinations by av_dlog. Update copyright year for ac3enc_opts_template.c. adts: Adjust frame size mask to follow the specification. APIchanges: fill hash for the avfilter_get_audio_buffer_ref_from_arrays addition lavfi: avfilter_merge_formats: handle case where inputs are same lavfi: use avfilter_get_audio_buffer_ref_from_arrays() in defaults.c lavfi: implement avfilter_get_audio_buffer_ref_from_arrays() APIchanges: remove duplicated entry APIchanges: fill in dates and numbers APIchanges: remove duplicated entry APIchanges: correctly interleave entries APIchanges: add entry for av_force_cpu_flags() addition lavf: bump minor after the addition of fps_probe_size to AVFormatContext lavc: bump minor after the addition of AVCodecContext.request_sample_fmt movenc: Add RTP muxer/hinter options movenc: Pass the RTP AVFormatContext to the SDP generation rtspenc: Add RTP muxer options rtspenc: Add an AVClass for setting muxer specific options rtpenc_chain: Pass the rtpflags options through to the chained muxer rtpenc: Declare the rtp flags private AVOptions in rtpenc.h sdp: Reindent after the previous commit rtpenc: MP4A-LATM payload support avoptions: Add an av_opt_flag_is_set function for inspecting flag fields sdp: Allow passing an AVFormatContext to the SDP generation mov: Fix wrong timestamp generation for fragmented movies that have time offset caused by the first edit list entry. mpeg12: more advanced ffmpeg mpeg2 aspect guessing code. ac3: there was no libav in 2010 thus this code cannot be from libav. swscale: split YUYV output out of yuv2packed[12X]_c(). dict: This code was developed in ffmpeg and not libav, nor by libav developers. Correct copyright notices. lavf: make compute_pkt_fields2() return meaningful error values matroskadec: set timestamps for RealAudio packets. intelh263dec: aspect ratio processing fix. intelh263dec: fix "Strict H.263 compliance" file playback oss,sndio: simplify by using FFMIN. swscale: extract monowhite/black output from yuv2packed[12X]_c(). swscale: de-macro'ify RGB15/16/32 input functions. swscale: rearrange code. movdec: Add support for the 'wfex' atom. ffmpeg.c: Add a necessary const qualifier riff: Fix potential memleak. swscale: change 48bit RGB input macros to inline functions. swscale: change 9/10bit YUV input macros to inline functions. swscale: extract gray16 output functions from yuv2packed[12X](). swscale: use standard clipping functions. swscale: merge macros that are used only once. swscale: fix function declarations in swscale.c. swscale: fix function declaration keywords in x86/swscale_template.c. jpegdec: actually search for and parse RSTn crypto: Use av_freep instead of av_free Revert "crypto: fix potential double free" Revert "build: remove empty $(OBJS) target" crypto: Use av_freep instead of av_free aac: fix adts frame size mask, fix demuxer probing for some files. lavf: don't try to free private options if priv_data is NULL. lavfi: handle NULL lists in avfilter_make_format_list swscale: fix types of assembly arguments. swscale: move two macros that are only used once into caller. swscale: remove unused function. Fix "mixed declarations and code" warnings. options: Add missing braces around struct initializer. mov: Remove leftover crufty debug statement with references to a local file. dvbsubdec: Fix compilation of debug code. Remove all uses of now deprecated metadata functions. Move metadata API from lavf to lavu. crypto: fix potential double free libx264: fix double free ffplay: remove -debug option ffplay: remove -vismv option mpegvideo: use av_get_picture_type_char() in ff_print_debug_info() Remove some non-compiling debug messages. ffplay: Fix non-compiling debug printf and replace it by av_dlog. H264: x86 predict init cosmetics. ac3enc: Fix linking of AC-3 encoder without the E-AC-3 encoder. Move E-AC-3 encoder functions to a separate eac3enc.c file. ac3enc: remove convenience macro, #define DEBUG ac3enc: remove unused #define vc1: re-initialize tables after width/height change. APIchanges: fill-in git commit hash for av_get_bytes_per_sample() addition samplefmt: add av_get_bytes_per_sample() libvpxenc: add forgotten AVClass. iirfilter: fix biquad filter coefficients. swscale: remove duplicate conversion routine in swScale(). swscale: add yuv2planar/packed function typedefs. swscale: integrate yuv2nv12X_C into yuv2yuvX() function pointers. swscale: reindent x86 init code. swscale: extract SWS_FULL_CHR_H_INT conditional into init code. swscale: cosmetics. swscale: remove alp/chr/lumSrcOffset. swscale: un-special-case yuv2yuvX16_c(). shorten: Remove stray DEBUG #define and corresponding av_dlog statement. vorbisdec: Restore mistakenly removed debug output. v4l2: set default standard to NULL sws: make dither_scale const configure: Document --enable-vdpau. Replace some av_log/printf + #ifdef combinations by av_dlog. Replace some nonstandard DEBUG_* preprocessor directives by plain DEBUG. svq1dec: Fix debug statements that referenced non-existing context. Replace some printf instances in debug code by av_log. showfiltfmts: use av_get_pix_fmt_name() inverse.c: Replace unnecessary intmath.h header by necessary stdint.h. Drop unnecessary directory prefixes from #include directives. Makefile: critical build fix after the merge. make fate passed locally due to ffmpeg/ffmpeg_g being there from before ffplay: Fix -vismv mem: Trying to workaround posix_memalign() bug on OSX build: remove empty $(OBJS) target build: make rule for linking ff* apply only to these targets eval: add support for pow() function build: rearrange some lines in a more logical way s302m: fix resampling for 16 and 24bits. ARM: remove MUL64 and MAC64 inline asm build: clean up .PHONY lists build: move all (un)install* target aliases to toplevel Makefile flvenc: propagate error properly build: remove stale dependency build: do not add CFLAGS-yes to CFLAGS utils.c: fix crash with threading enabled. configure: simplify source_path setup configure: remove --source-path option pixdesc: remove duplicated header inclusion lavfi: use av_samples_alloc() in avfilter_default_get_audio_buffer() lavfi: prefer nb_samples over size in AVFilterBufferRefAudioProps samplefmt: switch nb_channels/nb_samples params order in av_samples_alloc() samplefmt: change layout for arrays created by av_samples_alloc() and _fill_arrays() lavf: deprecate AVFormatParameters.time_base. img2: add framerate private option. img2: add video_size private option. img2: add pixel_format private option. tty: add framerate private option. Move code for "ffmpeg: fix massive leak occurring when seeking" / e4841a404bdabfeafb917454d510b60d888cb761 elsewhere lavf: remove reference to output-example in Makefile vsrc_buffer: add flags param to av_vsrc_buffer_add_video_buffer_ref Remove some unused scripts from tools/. Add x86 assembly for some 10-bit H.264 intra predict functions. v4l2: do not force NTSC as standard Add const to avfilter_get_video_buffer_ref_from_arrays arguments. Skip tableprint.h during 'make checkheaders'. Remove unnecessary LIBAVFORMAT_BUILD #ifdef. Drop explicit filenames from @file Doxygen tags. Skip generated table headers during 'make checkheaders'. lavf,lavc: free avoptions in a generic way. AVOptions: add av_opt_free convenience function. sdl: align option fields after last commit ffmpeg: fix massive leak occurring when seeking ffprobe: implement -i FILE option tableprint: Restore mistakenly deleted common.h #include for FF_ARRAY_ELEMS. ffplay.texi: document -i FILE option cmdutils: remove unnecessary OPT_DUMMY implementation cmdutils: change the signature of the function argument in parse_options() sdl: use the filename for defining the window title, if not specified tiff: print log in case of unknown / unsupported tag. tiff: fix linesize for mono-white/black formats. Fix build of eval-test program configure: Document --enable-vaapi swscale: override the lack of the accurate rounding flag when needed for dither. swscale: factor should_dither out ac3enc: extract all exponents for the frame at once ARM: remove MULL inline asm mathops: use MUL64 macro where it forms part of other ops tty: factorise returning error codes. rawdec: add framerate private option. x11grab: add framerate private option. fbdev,v4l2: remove some forgotten uses of AVFormatParameters.time_base. bktr: don't error when AVFormatParameters.time_base isn't set. cmdutils: add missing const qualifier Skip headers not designed to work standalone during 'make checkheaders'. Add missing #includes to make headers self-contained. musepack: remove unnecessary #include from mpcdata.h musepack: remove extraneous mpcdata.h inclusions Fix error check in av_file_map() udp: support old, crappy non pthread mode ffmpeg: use opt_acodec when setting audio codec in opt_target. ffmpeg: fix segfault with too many output files ffplay: error out with invalid sample rate or channels. oggdec: fix Ticket185 build: simplify commands for clean target j2kdec: dont fail on non zero cblock style. makefile: fix j2k encoder dependancies udp: fix indention swscale: split swscale.c in unscaled and generic conversion routines. swscale: cosmetics. swscale: integrate (literally) swscale_template.c in swscale.c. swscale: split out x86/swscale_template.c from swscale.c. swscale: enable hScale_altivec_real. swscale: split out ppc _template.c files from main swscale.c. swscale: remove indirections in ppc/swscale_template.c. swscale: split out unscaled altivec YUV converters in their own file. mpegvideoenc: fix multislice fate tests with threading disabled. cmdutils: move "#undef main" from ffplay.c to cmdutils.h wav: update size check for ds64 wav: fix skip size at end of ds64 chunk mpegts: Wrap #ifdef DEBUG and av_hex_dump_log() combination in a macro. build: Simplify texi2html invocation through the --output option. Mark some variables with av_unused Replace avcodec_get_pix_fmt_name() by av_get_pix_fmt_name(). svq3: Check negative mb_type to fix potential crash. svq3: Move svq3-specific fields to their own context. rawdec: initialize return value to 0. Remove unused get_psnr() prototype rawdec: don't leak option strings. bktr: get default framerate from video standard. swscale: remove unused COMPILE_TEMPLATE_ALTIVEC. h264 fill_filter_caches: Dont init chroma nnz_cache. In print_report, print progression time in hours:mins:secs:us ffmpeg: In print_report, use int64_t for pts to check for 0 and avoid inf value for bitrate. In libswscale, use all lines when converting from 422p to rgb with mmx, improve quality. Replace custom DEBUG preprocessor trickery by the standard one. vorbis: Remove non-compiling debug statement. vorbis: Remove pointless DEBUG #ifdef around debug output macros. cook: Remove non-compiling debug output. Remove pointless #ifdefs around function declarations in a header. Replace #ifdef + av_log() combinations by av_dlog(). Replace custom debug output functions by av_dlog(). cook: Remove unused debug functions. lavfi: add avfilter_link_free() function swscale: reintroduce sws_format_name() symbol Remove stray extra arguments from av_dlog() invocations. targa: fix big-endian build v4l2: remove one forgotten use of AVFormatParameters.pix_fmt. vfwcap: add a framerate private option. v4l2: add a framerate private option. libdc1394: add a framerate private option. fbdev: add a framerate private option. bktr: add a framerate private option. oma: check avio_read() return value nutdec: remove unused variable Remove unused variables swscale: dither for planar yuv outputs swscale: Fix use of uninitialized values (bug probably introduced from a marge of libav) cpudetect: add av_force_cpu_flags() swscale: allocate larger buffer to handle altivec overreads. H264/MPEG frame-level multi-threading. vsrc_buffer: propagate error code in av_vsrc_buffer_add_frame() lavfi: reindent after the previous commit lavfi: add braces around the block of an if() expression in avfilter_default_get_video_buffer lavfi: clarify the context of a comment in avfilter_default_get_video_buffer() lavfi: apply misc style fixes Cosmetic changes to h264_idct_10bit.asm. 2x faster h264_idct_add8_10. aacenc: Add stereo_mode option. h264: remove CONFIG_GPL from x86 intra prediction code. doc: cosmetics: libx264 typos postprocess: Remove test for impossible condition (was: Re: postprocess.c: replace check for p==NULL with *p==0) Fix various uninitialized variable warnings Port remove of get_sws_cpuflags from MPlayer's libmpcodecs. Replace "vector const" by "const vector" otherwise gcc 4.6.0 fails. Port recent changes to MPlayer libmpcodecs. Replace non-existent HAVE_SSE2 with HAVE_SSE. Simplify code and avoid compiler warning about incompatible types. Fix type of out[] variable, it should not be const. ARM: ac3dsp: optimised update_bap_counts() mpegaudiodec: Fix av_dlog() invocation. swscale: fix compilation of bfin due to missing pixdesc.h header lavf: tag dump_format() as @deprecated yuv4mpeg: complain and exit if a non-rawvideo stream is selected ffmpeg: handle copy of packets for AVFMT_RAWPICTURE output formats doc/examples: give meaningful names to the example files h264/10bit: add HAVE_ALIGNED_STACK checks. swscale: More accurate rounding in YSCALE_YUV_2_PACKEDX_FULL_C() Update 8-bit H.264 IDCT function names to reflect bit-depth. Add IDCT functions for 10-bit H.264. mpegaudioenc: Fix broken av_dlog statement. Employ correct printf format specifiers, mostly in debug output. ARM: fix MUL64 inline asm for pre-armv6 doc: add libvpx encoder section vf_drawtext: Replace FFmpeg by Libav in license boilerplate. mpegaudiodec: remove unusued code and variables postprocess.c: filter name needs to be double 0 terminated improved 'edts' atom writing support mpegaudio: clean up compute_antialias() definition vp8: fix segmentation race during frame-threading. Port libmpcodec fixes from MPlayer. Merge remote-tracking branch 'ffmpeg-mt/master' swscale: Remove unused variable. ARM: simplify inline asm with 64-bit operands Add "const" to avoid "initialization discards qualifiers" warning. Add const to fix "cast discards qualifiers" warnings. Include pixdesc.h for av_get_pix_fmt_name. wav: Don't avio_seek() if we know we'll run into EOF api-example: uppercase first letter in "copyright" output-example: create @file doxy from text in the copyright header examples: move API examples to a dedicated dir in doc ffmpeg: simplify opt_*_codec() options v4l2: rewrite code iterating the supported standards v4l2: perform minor style fixes v4l2: replace memset() with explicit struct initialization rawdec: fail in case of unknow pixel format swscale: remove sws_format_name() error.c: fix compile flags TCP: change default timeout to 5sec Revert "Timeout TCP open() after 5 seconds." Fix various unused variable warnings Fix various bad printf format warnings ARM: enable UAL syntax in asm.S Remove now unused nb_istreams variable. Add const to vector types for input in altivec code. Remove unused variable, avoiding compiler warning. Cast pointers to uintptr_t rather than unsigned int. v4l2: don't leak video standard string on error. swscale: Remove disabled code. avfilter: Surround function only used in debug mode by appropriate #ifdef. vf_crop: Replace #ifdef DEBUG + av_log() by av_dlog(). build: remove BUILD_ROOT variable vp8: use av_clip_uintp2() where possible swscale: Commits that could not be pulled earlier due to bugs #2 Commits that could not be pulled earlier due to bugs. Revert 1a5e4fd8c5b99478b4e08a69261930bb12aa948b for postproc. This broke the code doc: correct AC-3 option subsection placement ac3enc: fix LOCAL_ALIGNED usage in count_mantissa_bits() ac3dsp: do not use the ff_* prefix when referencing ff_ac3_bap_bits. swscale: use av_clip_uint8() in yuv2yuv1_c(). swscale: replace formatConvBuffer[VOF] by allocated array. v4l2: create file @doxy from text in the copyright header v4l2: remove pointless empty lines v4l2: set default standard to NULL v4l2: use OFFSET macro when setting options ac3dsp: fix loop condition in ac3_update_bap_counts_c() ARM: unbreak build lavdev: add SDL output device ac3enc: modify mantissa bit counting to keep bap counts for all values of bap instead of just 0 to 4. ac3enc: split mantissa bit counting into a separate function. ac3enc: store per-block/channel bap pointers by reference block in a 2D array rather than in the AC3Block struct. lavu: add av_get_pix_fmt_name() convenience function iff: remove duplicated file description cmdutils: remove OPT_FUNC2 get_bits: add av_unused tag to cache variable sws: replace all long with int. ARM: aacdec: fix constraints on inline asm ARM: remove unnecessary volatile from inline asm ARM: add "cc" clobbers to inline asm where needed ARM: improve FASTDIV asm ac3enc: use LOCAL_ALIGNED macro APIchanges: fill in git hash for av_get_pix_fmt_name (0420bd7). lavu: add av_get_pix_fmt_name() convenience function cmdutils: remove OPT_FUNC2 swscale: fix crash in bilinear scaling. vpxenc: add VP8E_SET_STATIC_THRESHOLD mapping webm: support stereo videos in matroska/webm muxer rgb2rgb: remove duplicate mmx/mmx2/3dnow/sse2 functions. swscale: reindent h[cy]scale_fast() and updateDitherTables(). swscale: reformat x86/swscale_template.c. swscale: remove duplicate mmx/mmx2 functions if they are identical. swscale: remove if (c->dstFormat) branch from yuv2packed[12X](). swscale: remove if(full_chr_int) from yuv2packed1(). swscale: remove if(accurate_rnd) branch from functions. swscale: revive SWS_CPU_CAPS until next major bump. swscale: Remove commented-out printf cruft. Export PCR pid Export more transport stream information. Output MPEG-TS stream identifiers. lavf: deprecate AVFormatParameters.pix_fmt. rawdec: add a pixel_format private option. v4l2: add a pixel_format private option. libdc1394: add a pixel_format private option. cosmetics: indentation and alignment after previous commit ac3enc: add support for E-AC-3 encoding. ac3enc: Move AC-3 AVOptions array to a separate file to make it easier to use only selected options for the different AC-3 encoder types. ARM: disable ff_vector_fmul_vfp on VFPv3 systems ARM: check for VFPv3 swscale: Remove unused variables in x86 code. doc: Drop DJGPP section, Libav now compiles out-of-the-box on FreeDOS. x86: Add appropriate ifdefs around certain AVX functions. cmdutils: use sws_freeContext() instead of av_freep(). swscale: delay allocation of formatConvBuffer(). swscale: fix build with --disable-swscale-alpha. movenc: Deprecate the global RTP hinting flag, use a private AVOption instead movenc: Add an AVClass for setting muxer specific options libdc1394: choose best video mode and rate based on camera capabilities. Remove support for libdc1394 < 2.0. avopt: fix segfault swscale: fix non-bitexact yuv2yuv[X2]() MMX/MMX2 functions. swscale: dont loose precission on RGB/BGR48 input, that is dont drop half the bits. patch checklist: suggest --disable-yasm test. lavdev: prefer the inclusion of avdevice.h over that of libavformat/avformat.h lavdev: include libavformat/avformat.h in avdevice.h fate.txt: replace FATE rsync command with a make command configure: report yasm/nasm presence properly tcp: make connect() timeout properly rawdec: factor video demuxer definitions into a macro. rtspdec: add initial_pause private option. lavf: deprecate AVFormatParameters.width/height. tty: add video_size private option. rawdec: add video_size private option. x11grab: add video_size private option. x11grab: factorize returning error codes. vfwcap: add video_size private option. v4l2: add video_size private option. v4l2: factorize returning error codes. libdc1394: add video_size private option. libdc1394: return meaninful error codes. bktr: add video_size private option. bktr: factorize returning error codes. Fix memleak Fix typo Remove specific note when not specific Minor cleanup in libx264.c Add metadata conversion table to the wav demuxer Fix 32bit rawvideo in avi on big-endian. id3v2: Check malloc result. ID3v2 tags can be very large. id3v2: Initialize tflags for version 2.2. webm: Additional options/presets for VP8 encodes under FFmpeg muxers: Add a flag to mark muxers that allow (non strict) monotone timestamps. swscale: Do not loose precission on yuv values after rgb->yuv. libx264: support aspect Ratio Switch ARM: add ARMv6 optimised av_clip_uintp2 ARM: remove volatile from asm statements in libavutil/intmath ARM: fix av_clipl_int32_arm() v4l: include avdevice.h ffserver: move close_connection() call to avoid a temporary string and copy. lavf: initialize demuxer private options. AVOptions: set string default values. Fix compilation with YASM/NASM versions not supporting AVX. lavdevice: mark v4l for removal on next major bump. swscale: fix compile on ppc. swscale: fix compile on x86-32. build: Remove generated .version file on distclean. configure: Add -D_GNU_SOURCE to CPPFLAGS on OS/2. doc: Drop hint at --enable-memalign-hack for MinGW, it is now autodetected. ffplay: Remove disabled code. Mark parameterless function declarations as 'void'. swscale: use av_clip_uint8() in yuv2yuv1_c(). swscale: remove VOF/VOFW. swscale: split chroma buffers into separate U/V planes. swscale: replace formatConvBuffer[VOF] by allocated array. rgb2rgb: remove duplicate mmx/mmx2/3dnow/sse2 functions. swscale: reindent h[cy]scale_fast() and updateDitherTables(). swscale: reformat x86/swscale_template.c. swscale: remove duplicate mmx/mmx2 functions if they are identical. swscale: remove if (c->dstFormat) branch from yuv2packed[12X](). swscale: remove if(full_chr_int) from yuv2packed1(). swscale: remove if(accurate_rnd) branch from functions. ffserver: Fix a null pointer dereference as a result of the FF_API_MAX_STREAMS cleanup. libdc1394: fix compilation. swscale: revive SWS_CPU_CAPS until next major bump. swscale: Remove commented-out printf cruft. ac3enc: initialize all coefficients to zero. ffv1: fix 16bits multithreading doc: create separate section for audio encoders swscale: Remove orphaned, commented-out function declaration. swscale: Eliminate rgb24toyv12_c() duplication. mpegvideo_enc: use AV_LOG_ERROR instead of AV_LOG_INFO for two error messages Fail when lowres value is lower than 0 Remove h263_msmpeg4 from MpegEncContext. APIchanges: Fill in git hash for fps_probe_size (30315a8) avformat: Add fpsprobesize as an AVOption. swscale: document SWS_CPU_CAPS* Revert removial of SWS flags from e66149e714006d099d1ebfcca3f22ca74fc7dcf4 avoptions: Return explicitly NAN or {0,0} if the option isn't found rtmp: Reindent rtmp: Don't try to do av_malloc(0) swscale: remove duplicatiopn of rgb24toyv12_c() Return -1 on invalid input instead of crashing. vf_mp: fix name of the remove-logo filter referenced in filters.texi tty: replace AVFormatParameters.sample_rate abuse with a private option. Fix end time of last chapter in compute_chapters_end ffmpeg: get rid of useless AVInputStream.nb_streams. ffmpeg: simplify managing input files and streams ffmpeg: purge redundant AVInputStream.index. lavf: deprecate AVFormatParameters.channel. libdc1394: add a private option for channel. dv1394: add a private option for channel. v4l2: reindent. v4l2: add a private option for channel. lavf: deprecate AVFormatParameters.standard. v4l2: add a private option for video standard. v4l: add a private option for video standard. dv1394: add a private option for video standard. bktr: add a private option for video standard. lavf: deprecate AVFormatParameters.{channels,sample_rate}. rawdec: add sample_rate/channels private options. ALSA: add channels and sample_rate private options. oss: add channels and sample_rate private options. sndio: add channels and sample_rate private options. lavf: deprecate AVFormatParameters.mpeg2ts_raw. mpegts: add compute_pcr option. lavf: add priv_class field to AVInputFormat. lavfi: add select filter eval: implement not() expression vsrc_buffer: return an error code if no frames are available ffmpeg: handle the case when get_filtered_frame() fails indeo3: add out-of-buffer write check Add reading of disc number to mov.c Fix end time of last chapter in compute_chapters_end(). Do not reset channel_layout to 0. vsrc_buffer: remove duplicated file description Merge swscale bloatup This will be cleaned up in the next merge swscale: MMX optim of hscale16() swscale: dont loose bits on planar >8bit yuv ind gray nput. swscale: Switch to ronalds yuv2yuvX16inC_template() its very similar to baptsites and supports alpha configure: enable memalign_hack automatically when needed rawdec: fix decoding of QT WRAW files matroska: improve declaration of video_stereo_* constant tables matroskadec: fix reverted condition to accept combine_plane operation Fix register types for LOAD_AB arguments, fixes compilation with NASM. swscale: unbreak the build on non-x86 systems. swscale: remove if(bitexact) branch from functions. swscale: remove if(canMMX2BeUsed) conditional. swscale: remove swScale_{c,MMX,MMX2} duplication. swscale: use emms_c(). Move emms_c() from libavcodec to libavutil. tiff: set palette in the context when specified in TIFF_PAL tag rtsp: use strtoul to parse rtptime and seq values. pgssubdec: fix incorrect colors. dvdsubdec: fix incorrect colors. ape: Allow demuxing of files with metadata tags. swscale: remove dead macro WRITEBGR24OLD. swscale: remove AMD3DNOW "optimizations". swscale: remove duplicate code in ppc/ subdirectory. swscale: remove duplicated x86/ functions. swscale: force --enable-runtime-cpudetect and remove SWS_CPU_CAPS_*. vsrc_buffer.h: add file doxy vsrc_buffer: tweak error message in init() wav: fix various printf warnings related to wrong argument type wav: propagate ff_get_wav_header() error code in w64_read_header() msmpeg4: reindent. lavc: remove msmpeg4v1 encoder. Remove avconfig.h and INCINSTDIRs on uninstall. ac3enc: add channel coupling support partial revert of 01d3ebaf219d83c0a70cdf9696ecb6b868e8a165 fate: reenable frext-pph10i4_panasonic_a after the bitstream has been fixed avcodec_find_decoder: prefer non experimental decoders. j2kdec: mark as CODEC_CAP_EXPERIMENTAL j2k[c/h] j2kdec.c: Implement 2 code block styles j2k: Add void as the parameter of function ff_j2k_init_tier1_luts Add Kamil Nowosads j2k code. matroska: cleanup handling of video stereo mode oggdec: use av_dlog() mem: define the MAX_MALLOC_SIZE constant and use it in place of INT_MAX configure: Add -U__STRICT_ANSI__ to CPPFLAGS on Cygwin and DOS. muxers.texi changes for mkv/webm options aacdec: fix typo in scalefactor clipping check mpegaudio: Correct license header add 5.1 to stereo downmix to resample.c this is based on previous 6to2channel-resample.patch from ffmpeg2theora but updated to work with trunk and using av_clip_int16. fate: fix fate-h264-conformance-frext-pph10i4-panasonic-a crcs. fate: update 9/10bit refs. h264: Properly set coded_{width, height} when parsing H.264. x86 asm: Add SECTION_TEXT to dct32_sse.asm. Fix 9/10 bit in swscale. Do not ask for samples if a specific channel layout was requested. libx264: specify field for default union values in options movdec: dont divide by zero when stts_data[0].duration = 0. Fix ticket127 dct32: Replacing libav by ffmpeg in the license header with the authors permission. Signed-off-by: Michael Niedermayer <michaelni@gmx.at> ffmpeg: Don't trigger url_interrupt_cb on the first signal avoptions: Check the return value from av_get_number lavf: fix style for avformat_alloc_output_context2() lavf: deprecate avformat_alloc_output_context() in favor of avformat_alloc_output_context2() lavfi: make vsrc_buffer.h header public dct32_sse: eliminate some spills Fix compilation with --disable-yasm. Fix dct32() compilation with --disable-yasm mpeg2dec: Fix lowres 3 lavfi: bump minor and add changelog entry after the split filter addition vf_split: add documentation to filters.texi vf_split: give more meaningful names to the output pads vf_split: define draw_slice() before end_frame() vf_split: add description vf_split: fix various nits wmadec: avoid infinit loop. DirectShow capture: Fix build ffmpeg: get rid of the -vglobal option. dct32: Add AVX implementation of 32-point DCT dct32: Change pass 6 permutation to allow for AVX implementation dct32: port SSE 32-point DCT to YASM matroska: switch stereo mode from int to string and add support in the demuxer too matroska: cosmetics Create a stereo_mode metadata tag to specify the stereo 3d video layout using the StereoMode tag in a matroska/webm video track. libavfilter: vf_split from soc. DirectShow capture support Signed-off-by: Michael Niedermayer <michaelni@gmx.at> multiple inclusion guard cleanup avio: document buffer must created with av_malloc() and friends avio: check AVIOContext malloc failure swscale: point out an alternative to sws_getContext svq3: Do initialization after parsing the extradata Fix channel_layout documentation. add changelog entries for 0.7_beta2 ffserver: dont just crash fix ffserver's SIGSEGV avoptions: Support getting flag values using av_get_int preset dir for win32 Merge remote-tracking branch 'ffmpeg-mt/master' Add a flag to disable side data merging. Merge/split side data. Encoding alac with more than two channels is not supported. mp3lame: add #include required for AV_RB32 macro. configure: make executable again LATM/AAC: Free previously initialized context on reinit. configure: Do not unconditionally add -Wall to host CFLAGS. configure: Set OS/2 objformat to a.out. Add support for a.out object format to assembler macros. fate: disable threading for encoding fate: add comment field fate: allow overriding default build and install dirs mpegtsenc: Add an AVClass pointer to the private data mpegaudio: clean up #includes mpegaudio: move all header parsing to mpegaudiodecheader.[ch] vf_libopencv: prefer opencv/cxcore.h over cxtypes.h decoders.texi: fix typos in rawvideo section cmdutils: use const AVClass * when senseful encoders.texi: add documentation for the libx264 encoder decoders.texi: add documentation for rawvideo decoder and options doc: add decoders.texi file encoders.texi: decrease level for audio encoders section ffprobe.texi: remove inclusion of muxers section indeo3: release buffer in indeo3_decode_end() indeo3: remove unnecessary includes indeo3: add @file doxy and a link to multimedia wiki documentation cmdutils: reset *picref_ptr to NULL in get_filtered_frame() ffmpeg: remove useless NULL-check on avfilter_unref_buffer libmp3lame: include "libavutil/intreadwrite.h" header qdm2: Use floating point synthesis filter. h264: correct border check. h264: fix loopfilter with threading at slice boundaries. Fix ff_mpa_synth_filter_fixed() prototype Reindent rtpenc_chain: Pass the MP4A_LATM flag to chained muxers rtpenc: MP4A-LATM payload support movenc: Pass AVFormatContext flags to the SDP generation sdp: Allow passing AVFormatContext flags to the SDP generation vsrc_buffer: document av_vsrc_buffer_add_video_buffer_ref() vsrc_buffer: add av_vsrc_buffer_add_frame() vsrc_buffer: fix example in docs, add mandatory parameters vsrc_buffer: make the source accept sws_param in init vsrc_buffer: propagate avfilter_open() error code vsrc_buffer: fix style lavfi: add avfilter_get_video_buffer_ref_from_frame to avcodec.h vsrc_buffer: remove dependency on AVFrame Rename costablegen.c ---> cos_tablegen.c. Collapse tableprint.c into tableprint.h. Simplify trig table rules Remove potentially unstable filenames from comments in generated files. Ignore generated tables and generated table generator programs. Simplify CLEANFILES make variable by using wildcards. Remove silly insults from avformat_version() Doxygen documentation. mpegaudiodsp: fix x86 and ppc makefiles configure: Adjust AVX assembler check. mpegaudio: remove unused version of SAME_HEADER_MASK mpegaudio: remove useless #undef at end of file asfdec: add missing #include for av_bswap32() mpegaudio: merge two #if CONFIG_FLOAT blocks mpegaudio: move some struct definitions from mpegaudio.h Move some mpegaudio functions to new mpegaudiodsp subsystem Clean up #includes in cmdutils.h. g729: Merge g729.h into g729dec.c. av_find_stream_info: Print more details about max anaylize duration failures. 10l: wrap float_interleave functions in HAVE_YASM. Add little description for -rc_override APIchanges: fill in date and commit for request_sample_fmt Add floating-point sample format support to the ac3, eac3, dca, aac, and vorbis decoders. Add support for request_sample_format in ffmpeg and ffplay. Add APIchanges entry for request_sample_fmt. Add request_sample_fmt field to AVCodecContext. Add float_interleave() to FmtConvertContext with x86-optimized versions. Remove unused make variable SEEK_REFFILE fate: remove redundant aref and vref references Parse 'bext' metadata in the wav demuxer Cosmetics: indent Keep parsing wav until EOF if the input is seekable and we know the size of the data tag Refactor the tag checking into a switch statement Use avio_tell() instead of url_ftell() add x264opts entry to docs cleaned up the udp.c, removed some variables and an av_log configure: favor pkg_config over sdl_config libx264: support passing arbitrary parameters. ffmpeg: dont show_banner() on verbose<0 fate: remove do_ffmpeg_nocheck function fate: do not collect -benchmark output mpegaudiodec: remove decode_end() function fate: run aref and vref as regular tests mpegaudio: sanitise compute_antialias_* names mpeg12: add slice-threading checks to slice-threading initializers. h264: copy pixel_shift between slice threading contexts. mdec: enable frame-level multithreading. mdec.c: fix overread. id3v2: prevent unsigned integer overflow in ff_id3v2_parse() id3v2: add @file doxy and link to format documentation configure: opensolaris install is not compatible with ffmpeg, allow overriding it. Fix compilation of iirfilter-test. eval: opensolaris strtod() cannot handle 0x1234 libx264: handle closed GOP codec flag lavf: remove duplicate assignment in avformat_alloc_context. lavf: use designated initializers for AVClasses. Make sure neither data_size nor sample_count is negative Refactor the 'fmt ' tag search and parsing flvdec: clenup debug code asfdec: fix possible overread on broken files. asfdec: do not fall back to binary/generic search asfdec: reindent after previous commit c7bd5ed asfdec: fallback to binary search internally mpegaudio: add _fixed suffix to some names Modify x86util.asm to ease transitioning to 10-bit H.264 assembly. ffmpeg: reset top_field_first in opt_input_file(). dct: build dct32 as separate object files qdm2: include correct header for rdft Ogg demuxer: give meaningful error codes and warnings. update changelog with 9/10 bit H264 and FFV1 changes Add some forgotten const to function arguments in libavfilter & libavformat. Write channel_layout for multichannel aif files. Fix ff_mov_write_chan() so it can be used by other muxers. Fix some mov files with little endian audio (tickets 201 - 203). iff/8svx: redesign 8SVX demuxing and decoding for handling stereo samples correctly iff: compact code setting metadata tags iff: fix bitrate computation for compressed audio stream iff: distinguish fields for audio and video compression imgutils: introduce internal image_get_linesize() and use it imgutils: make av_image_get_linesize() return AVERROR(EINVAL) for invalid pixel formats drawtext: specify union type for setting default options drawtext: reindent after the previous commit drawtext: fix strftime() text expansion ffmpeg: fix -aspect cli option Restructure video filter implementation in ffmpeg.c. ffplay: remove audio_write_get_buf_size() forward declaration lavfi: print key-frame and picture type information in ff_dlog_ref() mathops: remove ancient confusing comment rawdec: Allow overriding top field first. ffmpeg: initialize input_codec array earlier. cmdutils: Allocate private decoder context if its not allocated yet. cws2fws: Improve error message wording. tools: Check the return value of write(). mpegaudio: move OUT_FMT macro to mpegaudiodec.c mpegaudio: remove OUT_MIN/MAX macros Add missing #includes to mp3_header_(de)compress bsf dct: fix indentation dct: bypass table allocation for DCT_II of size 32 pngdec: relax condition for setting monoblack pixel format h264dsp_mmx: Add #ifdefs around some mmxext functions on x86_64. Remove unused header mpegaudio3.h. Support decoding of 1bpp rawvideo in avi (ticket 205). Support decoding of 2bpp rawvideo in avi (ticket 206). Bump minor after adding a caf muxer. configure: another try on fixing osx/mingw SDL aacdec: Use float instead of int16_t for ltp_state to avoid needless rounding. av_picture_crop(): Support simple cases with packed pixels too. acelp: Remove unused gray_decode table. dfa: Remove unused variable. configure: Include AVX availability in summary output. rawdec: propagate pict_type information to the output frame showinfo: replace "CRC" by "checksum" showinfo: fix vertical align nit showinfo: fix computation of Adler checksum imgutils: generalize linesize computation for bitstream formats configure: use same CPPFLAGS in kFreeBSD as Linux Conflicts: ffserver.c libavcodec/avcodec.h libavcodec/opt.h libavcodec/version.h libavdevice/avdevice.h libavfilter/avfilter.h libavformat/avformat.h libavformat/metadata.c libavformat/metadata.h libavformat/utils.c libavformat/version.h libavutil/avutil.h libavutil/mem.c Merged-by: Michael Niedermayer <michaelni@gmx.at>
3911 lines
143 KiB
C
3911 lines
143 KiB
C
/*
|
|
* VC-1 and WMV3 decoder
|
|
* Copyright (c) 2006-2007 Konstantin Shishkov
|
|
* Partly based on vc9.c (c) 2005 Anonymous, Alex Beregszaszi, Michael Niedermayer
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* VC-1 and WMV3 decoder
|
|
*
|
|
*/
|
|
#include "internal.h"
|
|
#include "dsputil.h"
|
|
#include "avcodec.h"
|
|
#include "mpegvideo.h"
|
|
#include "h263.h"
|
|
#include "vc1.h"
|
|
#include "vc1data.h"
|
|
#include "vc1acdata.h"
|
|
#include "msmpeg4data.h"
|
|
#include "unary.h"
|
|
#include "simple_idct.h"
|
|
#include "mathops.h"
|
|
#include "vdpau_internal.h"
|
|
|
|
#undef NDEBUG
|
|
#include <assert.h>
|
|
|
|
#define MB_INTRA_VLC_BITS 9
|
|
#define DC_VLC_BITS 9
|
|
#define AC_VLC_BITS 9
|
|
static const uint16_t table_mb_intra[64][2];
|
|
|
|
|
|
static const uint16_t vlc_offs[] = {
|
|
0, 520, 552, 616, 1128, 1160, 1224, 1740, 1772, 1836, 1900, 2436,
|
|
2986, 3050, 3610, 4154, 4218, 4746, 5326, 5390, 5902, 6554, 7658, 8620,
|
|
9262, 10202, 10756, 11310, 12228, 15078
|
|
};
|
|
|
|
/**
|
|
* Init VC-1 specific tables and VC1Context members
|
|
* @param v The VC1Context to initialize
|
|
* @return Status
|
|
*/
|
|
static int vc1_init_common(VC1Context *v)
|
|
{
|
|
static int done = 0;
|
|
int i = 0;
|
|
static VLC_TYPE vlc_table[15078][2];
|
|
|
|
v->hrd_rate = v->hrd_buffer = NULL;
|
|
|
|
/* VLC tables */
|
|
if(!done)
|
|
{
|
|
INIT_VLC_STATIC(&ff_vc1_bfraction_vlc, VC1_BFRACTION_VLC_BITS, 23,
|
|
ff_vc1_bfraction_bits, 1, 1,
|
|
ff_vc1_bfraction_codes, 1, 1, 1 << VC1_BFRACTION_VLC_BITS);
|
|
INIT_VLC_STATIC(&ff_vc1_norm2_vlc, VC1_NORM2_VLC_BITS, 4,
|
|
ff_vc1_norm2_bits, 1, 1,
|
|
ff_vc1_norm2_codes, 1, 1, 1 << VC1_NORM2_VLC_BITS);
|
|
INIT_VLC_STATIC(&ff_vc1_norm6_vlc, VC1_NORM6_VLC_BITS, 64,
|
|
ff_vc1_norm6_bits, 1, 1,
|
|
ff_vc1_norm6_codes, 2, 2, 556);
|
|
INIT_VLC_STATIC(&ff_vc1_imode_vlc, VC1_IMODE_VLC_BITS, 7,
|
|
ff_vc1_imode_bits, 1, 1,
|
|
ff_vc1_imode_codes, 1, 1, 1 << VC1_IMODE_VLC_BITS);
|
|
for (i=0; i<3; i++)
|
|
{
|
|
ff_vc1_ttmb_vlc[i].table = &vlc_table[vlc_offs[i*3+0]];
|
|
ff_vc1_ttmb_vlc[i].table_allocated = vlc_offs[i*3+1] - vlc_offs[i*3+0];
|
|
init_vlc(&ff_vc1_ttmb_vlc[i], VC1_TTMB_VLC_BITS, 16,
|
|
ff_vc1_ttmb_bits[i], 1, 1,
|
|
ff_vc1_ttmb_codes[i], 2, 2, INIT_VLC_USE_NEW_STATIC);
|
|
ff_vc1_ttblk_vlc[i].table = &vlc_table[vlc_offs[i*3+1]];
|
|
ff_vc1_ttblk_vlc[i].table_allocated = vlc_offs[i*3+2] - vlc_offs[i*3+1];
|
|
init_vlc(&ff_vc1_ttblk_vlc[i], VC1_TTBLK_VLC_BITS, 8,
|
|
ff_vc1_ttblk_bits[i], 1, 1,
|
|
ff_vc1_ttblk_codes[i], 1, 1, INIT_VLC_USE_NEW_STATIC);
|
|
ff_vc1_subblkpat_vlc[i].table = &vlc_table[vlc_offs[i*3+2]];
|
|
ff_vc1_subblkpat_vlc[i].table_allocated = vlc_offs[i*3+3] - vlc_offs[i*3+2];
|
|
init_vlc(&ff_vc1_subblkpat_vlc[i], VC1_SUBBLKPAT_VLC_BITS, 15,
|
|
ff_vc1_subblkpat_bits[i], 1, 1,
|
|
ff_vc1_subblkpat_codes[i], 1, 1, INIT_VLC_USE_NEW_STATIC);
|
|
}
|
|
for(i=0; i<4; i++)
|
|
{
|
|
ff_vc1_4mv_block_pattern_vlc[i].table = &vlc_table[vlc_offs[i*3+9]];
|
|
ff_vc1_4mv_block_pattern_vlc[i].table_allocated = vlc_offs[i*3+10] - vlc_offs[i*3+9];
|
|
init_vlc(&ff_vc1_4mv_block_pattern_vlc[i], VC1_4MV_BLOCK_PATTERN_VLC_BITS, 16,
|
|
ff_vc1_4mv_block_pattern_bits[i], 1, 1,
|
|
ff_vc1_4mv_block_pattern_codes[i], 1, 1, INIT_VLC_USE_NEW_STATIC);
|
|
ff_vc1_cbpcy_p_vlc[i].table = &vlc_table[vlc_offs[i*3+10]];
|
|
ff_vc1_cbpcy_p_vlc[i].table_allocated = vlc_offs[i*3+11] - vlc_offs[i*3+10];
|
|
init_vlc(&ff_vc1_cbpcy_p_vlc[i], VC1_CBPCY_P_VLC_BITS, 64,
|
|
ff_vc1_cbpcy_p_bits[i], 1, 1,
|
|
ff_vc1_cbpcy_p_codes[i], 2, 2, INIT_VLC_USE_NEW_STATIC);
|
|
ff_vc1_mv_diff_vlc[i].table = &vlc_table[vlc_offs[i*3+11]];
|
|
ff_vc1_mv_diff_vlc[i].table_allocated = vlc_offs[i*3+12] - vlc_offs[i*3+11];
|
|
init_vlc(&ff_vc1_mv_diff_vlc[i], VC1_MV_DIFF_VLC_BITS, 73,
|
|
ff_vc1_mv_diff_bits[i], 1, 1,
|
|
ff_vc1_mv_diff_codes[i], 2, 2, INIT_VLC_USE_NEW_STATIC);
|
|
}
|
|
for(i=0; i<8; i++){
|
|
ff_vc1_ac_coeff_table[i].table = &vlc_table[vlc_offs[i+21]];
|
|
ff_vc1_ac_coeff_table[i].table_allocated = vlc_offs[i+22] - vlc_offs[i+21];
|
|
init_vlc(&ff_vc1_ac_coeff_table[i], AC_VLC_BITS, vc1_ac_sizes[i],
|
|
&vc1_ac_tables[i][0][1], 8, 4,
|
|
&vc1_ac_tables[i][0][0], 8, 4, INIT_VLC_USE_NEW_STATIC);
|
|
}
|
|
done = 1;
|
|
}
|
|
|
|
/* Other defaults */
|
|
v->pq = -1;
|
|
v->mvrange = 0; /* 7.1.1.18, p80 */
|
|
|
|
return 0;
|
|
}
|
|
|
|
/***********************************************************************/
|
|
/**
|
|
* @defgroup vc1bitplane VC-1 Bitplane decoding
|
|
* @see 8.7, p56
|
|
* @{
|
|
*/
|
|
|
|
/**
|
|
* Imode types
|
|
* @{
|
|
*/
|
|
enum Imode {
|
|
IMODE_RAW,
|
|
IMODE_NORM2,
|
|
IMODE_DIFF2,
|
|
IMODE_NORM6,
|
|
IMODE_DIFF6,
|
|
IMODE_ROWSKIP,
|
|
IMODE_COLSKIP
|
|
};
|
|
/** @} */ //imode defines
|
|
|
|
|
|
/** @} */ //Bitplane group
|
|
|
|
static void vc1_put_signed_blocks_clamped(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
|
|
/* The put pixels loop is always one MB row behind the decoding loop,
|
|
* because we can only put pixels when overlap filtering is done, and
|
|
* for filtering of the bottom edge of a MB, we need the next MB row
|
|
* present as well.
|
|
* Within the row, the put pixels loop is also one MB col behind the
|
|
* decoding loop. The reason for this is again, because for filtering
|
|
* of the right MB edge, we need the next MB present. */
|
|
if (!s->first_slice_line) {
|
|
if (s->mb_x) {
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][0],
|
|
s->dest[0] - 16 * s->linesize - 16,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][1],
|
|
s->dest[0] - 16 * s->linesize - 8,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][2],
|
|
s->dest[0] - 8 * s->linesize - 16,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][3],
|
|
s->dest[0] - 8 * s->linesize - 8,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][4],
|
|
s->dest[1] - 8 * s->uvlinesize - 8,
|
|
s->uvlinesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->topleft_blk_idx][5],
|
|
s->dest[2] - 8 * s->uvlinesize - 8,
|
|
s->uvlinesize);
|
|
}
|
|
if (s->mb_x == s->mb_width - 1) {
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][0],
|
|
s->dest[0] - 16 * s->linesize,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][1],
|
|
s->dest[0] - 16 * s->linesize + 8,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][2],
|
|
s->dest[0] - 8 * s->linesize,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][3],
|
|
s->dest[0] - 8 * s->linesize + 8,
|
|
s->linesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][4],
|
|
s->dest[1] - 8 * s->uvlinesize,
|
|
s->uvlinesize);
|
|
s->dsp.put_signed_pixels_clamped(v->block[v->top_blk_idx][5],
|
|
s->dest[2] - 8 * s->uvlinesize,
|
|
s->uvlinesize);
|
|
}
|
|
}
|
|
|
|
#define inc_blk_idx(idx) do { \
|
|
idx++; \
|
|
if (idx >= v->n_allocated_blks) \
|
|
idx = 0; \
|
|
} while (0)
|
|
|
|
inc_blk_idx(v->topleft_blk_idx);
|
|
inc_blk_idx(v->top_blk_idx);
|
|
inc_blk_idx(v->left_blk_idx);
|
|
inc_blk_idx(v->cur_blk_idx);
|
|
}
|
|
|
|
static void vc1_loop_filter_iblk(VC1Context *v, int pq)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int j;
|
|
if (!s->first_slice_line) {
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0], s->linesize, pq);
|
|
if (s->mb_x)
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16*s->linesize, s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16*s->linesize+8, s->linesize, pq);
|
|
for(j = 0; j < 2; j++){
|
|
v->vc1dsp.vc1_v_loop_filter8(s->dest[j+1], s->uvlinesize, pq);
|
|
if (s->mb_x)
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[j+1]-8*s->uvlinesize, s->uvlinesize, pq);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0] + 8*s->linesize, s->linesize, pq);
|
|
|
|
if (s->mb_y == s->mb_height-1) {
|
|
if (s->mb_x) {
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0], s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[1], s->uvlinesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[2], s->uvlinesize, pq);
|
|
}
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] + 8, s->linesize, pq);
|
|
}
|
|
}
|
|
|
|
static void vc1_loop_filter_iblk_delayed(VC1Context *v, int pq)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int j;
|
|
|
|
/* The loopfilter runs 1 row and 1 column behind the overlap filter, which
|
|
* means it runs two rows/cols behind the decoding loop. */
|
|
if (!s->first_slice_line) {
|
|
if (s->mb_x) {
|
|
if (s->mb_y >= s->start_mb_y + 2) {
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0] - 16 * s->linesize - 16, s->linesize, pq);
|
|
|
|
if (s->mb_x >= 2)
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 32 * s->linesize - 16, s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 32 * s->linesize - 8, s->linesize, pq);
|
|
for(j = 0; j < 2; j++) {
|
|
v->vc1dsp.vc1_v_loop_filter8(s->dest[j+1] - 8 * s->uvlinesize - 8, s->uvlinesize, pq);
|
|
if (s->mb_x >= 2) {
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[j+1] - 16 * s->uvlinesize - 8, s->uvlinesize, pq);
|
|
}
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0] - 8 * s->linesize - 16, s->linesize, pq);
|
|
}
|
|
|
|
if (s->mb_x == s->mb_width - 1) {
|
|
if (s->mb_y >= s->start_mb_y + 2) {
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0] - 16 * s->linesize, s->linesize, pq);
|
|
|
|
if (s->mb_x)
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 32 * s->linesize, s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 32 * s->linesize + 8, s->linesize, pq);
|
|
for(j = 0; j < 2; j++) {
|
|
v->vc1dsp.vc1_v_loop_filter8(s->dest[j+1] - 8 * s->uvlinesize, s->uvlinesize, pq);
|
|
if (s->mb_x >= 2) {
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[j+1] - 16 * s->uvlinesize, s->uvlinesize, pq);
|
|
}
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_loop_filter16(s->dest[0] - 8 * s->linesize, s->linesize, pq);
|
|
}
|
|
|
|
if (s->mb_y == s->mb_height) {
|
|
if (s->mb_x) {
|
|
if (s->mb_x >= 2)
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16 * s->linesize - 16, s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16 * s->linesize - 8, s->linesize, pq);
|
|
if (s->mb_x >= 2) {
|
|
for(j = 0; j < 2; j++) {
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[j+1] - 8 * s->uvlinesize - 8, s->uvlinesize, pq);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (s->mb_x == s->mb_width - 1) {
|
|
if (s->mb_x)
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16 * s->linesize, s->linesize, pq);
|
|
v->vc1dsp.vc1_h_loop_filter16(s->dest[0] - 16 * s->linesize + 8, s->linesize, pq);
|
|
if (s->mb_x) {
|
|
for(j = 0; j < 2; j++) {
|
|
v->vc1dsp.vc1_h_loop_filter8(s->dest[j+1] - 8 * s->uvlinesize, s->uvlinesize, pq);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
static void vc1_smooth_overlap_filter_iblk(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int mb_pos;
|
|
|
|
if (v->condover == CONDOVER_NONE)
|
|
return;
|
|
|
|
mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
|
|
/* Within a MB, the horizontal overlap always runs before the vertical.
|
|
* To accomplish that, we run the H on left and internal borders of the
|
|
* currently decoded MB. Then, we wait for the next overlap iteration
|
|
* to do H overlap on the right edge of this MB, before moving over and
|
|
* running the V overlap. Therefore, the V overlap makes us trail by one
|
|
* MB col and the H overlap filter makes us trail by one MB row. This
|
|
* is reflected in the time at which we run the put_pixels loop. */
|
|
if(v->condover == CONDOVER_ALL || v->pq >= 9 || v->over_flags_plane[mb_pos]) {
|
|
if(s->mb_x && (v->condover == CONDOVER_ALL || v->pq >= 9 ||
|
|
v->over_flags_plane[mb_pos - 1])) {
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->left_blk_idx][1],
|
|
v->block[v->cur_blk_idx][0]);
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->left_blk_idx][3],
|
|
v->block[v->cur_blk_idx][2]);
|
|
if(!(s->flags & CODEC_FLAG_GRAY)) {
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->left_blk_idx][4],
|
|
v->block[v->cur_blk_idx][4]);
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->left_blk_idx][5],
|
|
v->block[v->cur_blk_idx][5]);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->cur_blk_idx][0],
|
|
v->block[v->cur_blk_idx][1]);
|
|
v->vc1dsp.vc1_h_s_overlap(v->block[v->cur_blk_idx][2],
|
|
v->block[v->cur_blk_idx][3]);
|
|
|
|
if (s->mb_x == s->mb_width - 1) {
|
|
if(!s->first_slice_line && (v->condover == CONDOVER_ALL || v->pq >= 9 ||
|
|
v->over_flags_plane[mb_pos - s->mb_stride])) {
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->top_blk_idx][2],
|
|
v->block[v->cur_blk_idx][0]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->top_blk_idx][3],
|
|
v->block[v->cur_blk_idx][1]);
|
|
if(!(s->flags & CODEC_FLAG_GRAY)) {
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->top_blk_idx][4],
|
|
v->block[v->cur_blk_idx][4]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->top_blk_idx][5],
|
|
v->block[v->cur_blk_idx][5]);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->cur_blk_idx][0],
|
|
v->block[v->cur_blk_idx][2]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->cur_blk_idx][1],
|
|
v->block[v->cur_blk_idx][3]);
|
|
}
|
|
}
|
|
if (s->mb_x && (v->condover == CONDOVER_ALL || v->over_flags_plane[mb_pos - 1])) {
|
|
if(!s->first_slice_line && (v->condover == CONDOVER_ALL || v->pq >= 9 ||
|
|
v->over_flags_plane[mb_pos - s->mb_stride - 1])) {
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->topleft_blk_idx][2],
|
|
v->block[v->left_blk_idx][0]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->topleft_blk_idx][3],
|
|
v->block[v->left_blk_idx][1]);
|
|
if(!(s->flags & CODEC_FLAG_GRAY)) {
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->topleft_blk_idx][4],
|
|
v->block[v->left_blk_idx][4]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->topleft_blk_idx][5],
|
|
v->block[v->left_blk_idx][5]);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->left_blk_idx][0],
|
|
v->block[v->left_blk_idx][2]);
|
|
v->vc1dsp.vc1_v_s_overlap(v->block[v->left_blk_idx][1],
|
|
v->block[v->left_blk_idx][3]);
|
|
}
|
|
}
|
|
|
|
/** Do motion compensation over 1 macroblock
|
|
* Mostly adapted hpel_motion and qpel_motion from mpegvideo.c
|
|
*/
|
|
static void vc1_mc_1mv(VC1Context *v, int dir)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
DSPContext *dsp = &v->s.dsp;
|
|
uint8_t *srcY, *srcU, *srcV;
|
|
int dxy, mx, my, uvmx, uvmy, src_x, src_y, uvsrc_x, uvsrc_y;
|
|
|
|
if(!v->s.last_picture.data[0])return;
|
|
|
|
mx = s->mv[dir][0][0];
|
|
my = s->mv[dir][0][1];
|
|
|
|
// store motion vectors for further use in B frames
|
|
if(s->pict_type == AV_PICTURE_TYPE_P) {
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = mx;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = my;
|
|
}
|
|
uvmx = (mx + ((mx & 3) == 3)) >> 1;
|
|
uvmy = (my + ((my & 3) == 3)) >> 1;
|
|
v->luma_mv[s->mb_x][0] = uvmx;
|
|
v->luma_mv[s->mb_x][1] = uvmy;
|
|
if(v->fastuvmc) {
|
|
uvmx = uvmx + ((uvmx<0)?(uvmx&1):-(uvmx&1));
|
|
uvmy = uvmy + ((uvmy<0)?(uvmy&1):-(uvmy&1));
|
|
}
|
|
if(!dir) {
|
|
srcY = s->last_picture.data[0];
|
|
srcU = s->last_picture.data[1];
|
|
srcV = s->last_picture.data[2];
|
|
} else {
|
|
srcY = s->next_picture.data[0];
|
|
srcU = s->next_picture.data[1];
|
|
srcV = s->next_picture.data[2];
|
|
}
|
|
|
|
src_x = s->mb_x * 16 + (mx >> 2);
|
|
src_y = s->mb_y * 16 + (my >> 2);
|
|
uvsrc_x = s->mb_x * 8 + (uvmx >> 2);
|
|
uvsrc_y = s->mb_y * 8 + (uvmy >> 2);
|
|
|
|
if(v->profile != PROFILE_ADVANCED){
|
|
src_x = av_clip( src_x, -16, s->mb_width * 16);
|
|
src_y = av_clip( src_y, -16, s->mb_height * 16);
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->mb_width * 8);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->mb_height * 8);
|
|
}else{
|
|
src_x = av_clip( src_x, -17, s->avctx->coded_width);
|
|
src_y = av_clip( src_y, -18, s->avctx->coded_height + 1);
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->avctx->coded_width >> 1);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->avctx->coded_height >> 1);
|
|
}
|
|
|
|
srcY += src_y * s->linesize + src_x;
|
|
srcU += uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
srcV += uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
|
|
/* for grayscale we should not try to read from unknown area */
|
|
if(s->flags & CODEC_FLAG_GRAY) {
|
|
srcU = s->edge_emu_buffer + 18 * s->linesize;
|
|
srcV = s->edge_emu_buffer + 18 * s->linesize;
|
|
}
|
|
|
|
if(v->rangeredfrm || (v->mv_mode == MV_PMODE_INTENSITY_COMP)
|
|
|| (unsigned)(src_x - s->mspel) > s->h_edge_pos - (mx&3) - 16 - s->mspel*3
|
|
|| (unsigned)(src_y - s->mspel) > s->v_edge_pos - (my&3) - 16 - s->mspel*3){
|
|
uint8_t *uvbuf= s->edge_emu_buffer + 19 * s->linesize;
|
|
|
|
srcY -= s->mspel * (1 + s->linesize);
|
|
s->dsp.emulated_edge_mc(s->edge_emu_buffer, srcY, s->linesize, 17+s->mspel*2, 17+s->mspel*2,
|
|
src_x - s->mspel, src_y - s->mspel, s->h_edge_pos, s->v_edge_pos);
|
|
srcY = s->edge_emu_buffer;
|
|
s->dsp.emulated_edge_mc(uvbuf , srcU, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
s->dsp.emulated_edge_mc(uvbuf + 16, srcV, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
srcU = uvbuf;
|
|
srcV = uvbuf + 16;
|
|
/* if we deal with range reduction we need to scale source blocks */
|
|
if(v->rangeredfrm) {
|
|
int i, j;
|
|
uint8_t *src, *src2;
|
|
|
|
src = srcY;
|
|
for(j = 0; j < 17 + s->mspel*2; j++) {
|
|
for(i = 0; i < 17 + s->mspel*2; i++) src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src += s->linesize;
|
|
}
|
|
src = srcU; src2 = srcV;
|
|
for(j = 0; j < 9; j++) {
|
|
for(i = 0; i < 9; i++) {
|
|
src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src2[i] = ((src2[i] - 128) >> 1) + 128;
|
|
}
|
|
src += s->uvlinesize;
|
|
src2 += s->uvlinesize;
|
|
}
|
|
}
|
|
/* if we deal with intensity compensation we need to scale source blocks */
|
|
if(v->mv_mode == MV_PMODE_INTENSITY_COMP) {
|
|
int i, j;
|
|
uint8_t *src, *src2;
|
|
|
|
src = srcY;
|
|
for(j = 0; j < 17 + s->mspel*2; j++) {
|
|
for(i = 0; i < 17 + s->mspel*2; i++) src[i] = v->luty[src[i]];
|
|
src += s->linesize;
|
|
}
|
|
src = srcU; src2 = srcV;
|
|
for(j = 0; j < 9; j++) {
|
|
for(i = 0; i < 9; i++) {
|
|
src[i] = v->lutuv[src[i]];
|
|
src2[i] = v->lutuv[src2[i]];
|
|
}
|
|
src += s->uvlinesize;
|
|
src2 += s->uvlinesize;
|
|
}
|
|
}
|
|
srcY += s->mspel * (1 + s->linesize);
|
|
}
|
|
|
|
if(s->mspel) {
|
|
dxy = ((my & 3) << 2) | (mx & 3);
|
|
v->vc1dsp.put_vc1_mspel_pixels_tab[dxy](s->dest[0] , srcY , s->linesize, v->rnd);
|
|
v->vc1dsp.put_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8, srcY + 8, s->linesize, v->rnd);
|
|
srcY += s->linesize * 8;
|
|
v->vc1dsp.put_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8 * s->linesize , srcY , s->linesize, v->rnd);
|
|
v->vc1dsp.put_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8 * s->linesize + 8, srcY + 8, s->linesize, v->rnd);
|
|
} else { // hpel mc - always used for luma
|
|
dxy = (my & 2) | ((mx & 2) >> 1);
|
|
|
|
if(!v->rnd)
|
|
dsp->put_pixels_tab[0][dxy](s->dest[0], srcY, s->linesize, 16);
|
|
else
|
|
dsp->put_no_rnd_pixels_tab[0][dxy](s->dest[0], srcY, s->linesize, 16);
|
|
}
|
|
|
|
if(s->flags & CODEC_FLAG_GRAY) return;
|
|
/* Chroma MC always uses qpel bilinear */
|
|
uvmx = (uvmx&3)<<1;
|
|
uvmy = (uvmy&3)<<1;
|
|
if(!v->rnd){
|
|
dsp->put_h264_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
dsp->put_h264_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}else{
|
|
v->vc1dsp.put_no_rnd_vc1_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
v->vc1dsp.put_no_rnd_vc1_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}
|
|
}
|
|
|
|
/** Do motion compensation for 4-MV macroblock - luminance block
|
|
*/
|
|
static void vc1_mc_4mv_luma(VC1Context *v, int n)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
DSPContext *dsp = &v->s.dsp;
|
|
uint8_t *srcY;
|
|
int dxy, mx, my, src_x, src_y;
|
|
int off;
|
|
|
|
if(!v->s.last_picture.data[0])return;
|
|
mx = s->mv[0][n][0];
|
|
my = s->mv[0][n][1];
|
|
srcY = s->last_picture.data[0];
|
|
|
|
off = s->linesize * 4 * (n&2) + (n&1) * 8;
|
|
|
|
src_x = s->mb_x * 16 + (n&1) * 8 + (mx >> 2);
|
|
src_y = s->mb_y * 16 + (n&2) * 4 + (my >> 2);
|
|
|
|
if(v->profile != PROFILE_ADVANCED){
|
|
src_x = av_clip( src_x, -16, s->mb_width * 16);
|
|
src_y = av_clip( src_y, -16, s->mb_height * 16);
|
|
}else{
|
|
src_x = av_clip( src_x, -17, s->avctx->coded_width);
|
|
src_y = av_clip( src_y, -18, s->avctx->coded_height + 1);
|
|
}
|
|
|
|
srcY += src_y * s->linesize + src_x;
|
|
|
|
if(v->rangeredfrm || (v->mv_mode == MV_PMODE_INTENSITY_COMP)
|
|
|| (unsigned)(src_x - s->mspel) > s->h_edge_pos - (mx&3) - 8 - s->mspel*2
|
|
|| (unsigned)(src_y - s->mspel) > s->v_edge_pos - (my&3) - 8 - s->mspel*2){
|
|
srcY -= s->mspel * (1 + s->linesize);
|
|
s->dsp.emulated_edge_mc(s->edge_emu_buffer, srcY, s->linesize, 9+s->mspel*2, 9+s->mspel*2,
|
|
src_x - s->mspel, src_y - s->mspel, s->h_edge_pos, s->v_edge_pos);
|
|
srcY = s->edge_emu_buffer;
|
|
/* if we deal with range reduction we need to scale source blocks */
|
|
if(v->rangeredfrm) {
|
|
int i, j;
|
|
uint8_t *src;
|
|
|
|
src = srcY;
|
|
for(j = 0; j < 9 + s->mspel*2; j++) {
|
|
for(i = 0; i < 9 + s->mspel*2; i++) src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src += s->linesize;
|
|
}
|
|
}
|
|
/* if we deal with intensity compensation we need to scale source blocks */
|
|
if(v->mv_mode == MV_PMODE_INTENSITY_COMP) {
|
|
int i, j;
|
|
uint8_t *src;
|
|
|
|
src = srcY;
|
|
for(j = 0; j < 9 + s->mspel*2; j++) {
|
|
for(i = 0; i < 9 + s->mspel*2; i++) src[i] = v->luty[src[i]];
|
|
src += s->linesize;
|
|
}
|
|
}
|
|
srcY += s->mspel * (1 + s->linesize);
|
|
}
|
|
|
|
if(s->mspel) {
|
|
dxy = ((my & 3) << 2) | (mx & 3);
|
|
v->vc1dsp.put_vc1_mspel_pixels_tab[dxy](s->dest[0] + off, srcY, s->linesize, v->rnd);
|
|
} else { // hpel mc - always used for luma
|
|
dxy = (my & 2) | ((mx & 2) >> 1);
|
|
if(!v->rnd)
|
|
dsp->put_pixels_tab[1][dxy](s->dest[0] + off, srcY, s->linesize, 8);
|
|
else
|
|
dsp->put_no_rnd_pixels_tab[1][dxy](s->dest[0] + off, srcY, s->linesize, 8);
|
|
}
|
|
}
|
|
|
|
static inline int median4(int a, int b, int c, int d)
|
|
{
|
|
if(a < b) {
|
|
if(c < d) return (FFMIN(b, d) + FFMAX(a, c)) / 2;
|
|
else return (FFMIN(b, c) + FFMAX(a, d)) / 2;
|
|
} else {
|
|
if(c < d) return (FFMIN(a, d) + FFMAX(b, c)) / 2;
|
|
else return (FFMIN(a, c) + FFMAX(b, d)) / 2;
|
|
}
|
|
}
|
|
|
|
|
|
/** Do motion compensation for 4-MV macroblock - both chroma blocks
|
|
*/
|
|
static void vc1_mc_4mv_chroma(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
DSPContext *dsp = &v->s.dsp;
|
|
uint8_t *srcU, *srcV;
|
|
int uvmx, uvmy, uvsrc_x, uvsrc_y;
|
|
int i, idx, tx = 0, ty = 0;
|
|
int mvx[4], mvy[4], intra[4];
|
|
static const int count[16] = { 0, 1, 1, 2, 1, 2, 2, 3, 1, 2, 2, 3, 2, 3, 3, 4};
|
|
|
|
if(!v->s.last_picture.data[0])return;
|
|
if(s->flags & CODEC_FLAG_GRAY) return;
|
|
|
|
for(i = 0; i < 4; i++) {
|
|
mvx[i] = s->mv[0][i][0];
|
|
mvy[i] = s->mv[0][i][1];
|
|
intra[i] = v->mb_type[0][s->block_index[i]];
|
|
}
|
|
|
|
/* calculate chroma MV vector from four luma MVs */
|
|
idx = (intra[3] << 3) | (intra[2] << 2) | (intra[1] << 1) | intra[0];
|
|
if(!idx) { // all blocks are inter
|
|
tx = median4(mvx[0], mvx[1], mvx[2], mvx[3]);
|
|
ty = median4(mvy[0], mvy[1], mvy[2], mvy[3]);
|
|
} else if(count[idx] == 1) { // 3 inter blocks
|
|
switch(idx) {
|
|
case 0x1:
|
|
tx = mid_pred(mvx[1], mvx[2], mvx[3]);
|
|
ty = mid_pred(mvy[1], mvy[2], mvy[3]);
|
|
break;
|
|
case 0x2:
|
|
tx = mid_pred(mvx[0], mvx[2], mvx[3]);
|
|
ty = mid_pred(mvy[0], mvy[2], mvy[3]);
|
|
break;
|
|
case 0x4:
|
|
tx = mid_pred(mvx[0], mvx[1], mvx[3]);
|
|
ty = mid_pred(mvy[0], mvy[1], mvy[3]);
|
|
break;
|
|
case 0x8:
|
|
tx = mid_pred(mvx[0], mvx[1], mvx[2]);
|
|
ty = mid_pred(mvy[0], mvy[1], mvy[2]);
|
|
break;
|
|
}
|
|
} else if(count[idx] == 2) {
|
|
int t1 = 0, t2 = 0;
|
|
for(i=0; i<3;i++) if(!intra[i]) {t1 = i; break;}
|
|
for(i= t1+1; i<4; i++)if(!intra[i]) {t2 = i; break;}
|
|
tx = (mvx[t1] + mvx[t2]) / 2;
|
|
ty = (mvy[t1] + mvy[t2]) / 2;
|
|
} else {
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = 0;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = 0;
|
|
v->luma_mv[s->mb_x][0] = v->luma_mv[s->mb_x][1] = 0;
|
|
return; //no need to do MC for inter blocks
|
|
}
|
|
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = tx;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = ty;
|
|
uvmx = (tx + ((tx&3) == 3)) >> 1;
|
|
uvmy = (ty + ((ty&3) == 3)) >> 1;
|
|
v->luma_mv[s->mb_x][0] = uvmx;
|
|
v->luma_mv[s->mb_x][1] = uvmy;
|
|
if(v->fastuvmc) {
|
|
uvmx = uvmx + ((uvmx<0)?(uvmx&1):-(uvmx&1));
|
|
uvmy = uvmy + ((uvmy<0)?(uvmy&1):-(uvmy&1));
|
|
}
|
|
|
|
uvsrc_x = s->mb_x * 8 + (uvmx >> 2);
|
|
uvsrc_y = s->mb_y * 8 + (uvmy >> 2);
|
|
|
|
if(v->profile != PROFILE_ADVANCED){
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->mb_width * 8);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->mb_height * 8);
|
|
}else{
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->avctx->coded_width >> 1);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->avctx->coded_height >> 1);
|
|
}
|
|
|
|
srcU = s->last_picture.data[1] + uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
srcV = s->last_picture.data[2] + uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
if(v->rangeredfrm || (v->mv_mode == MV_PMODE_INTENSITY_COMP)
|
|
|| (unsigned)uvsrc_x > (s->h_edge_pos >> 1) - 9
|
|
|| (unsigned)uvsrc_y > (s->v_edge_pos >> 1) - 9){
|
|
s->dsp.emulated_edge_mc(s->edge_emu_buffer , srcU, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
s->dsp.emulated_edge_mc(s->edge_emu_buffer + 16, srcV, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
srcU = s->edge_emu_buffer;
|
|
srcV = s->edge_emu_buffer + 16;
|
|
|
|
/* if we deal with range reduction we need to scale source blocks */
|
|
if(v->rangeredfrm) {
|
|
int i, j;
|
|
uint8_t *src, *src2;
|
|
|
|
src = srcU; src2 = srcV;
|
|
for(j = 0; j < 9; j++) {
|
|
for(i = 0; i < 9; i++) {
|
|
src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src2[i] = ((src2[i] - 128) >> 1) + 128;
|
|
}
|
|
src += s->uvlinesize;
|
|
src2 += s->uvlinesize;
|
|
}
|
|
}
|
|
/* if we deal with intensity compensation we need to scale source blocks */
|
|
if(v->mv_mode == MV_PMODE_INTENSITY_COMP) {
|
|
int i, j;
|
|
uint8_t *src, *src2;
|
|
|
|
src = srcU; src2 = srcV;
|
|
for(j = 0; j < 9; j++) {
|
|
for(i = 0; i < 9; i++) {
|
|
src[i] = v->lutuv[src[i]];
|
|
src2[i] = v->lutuv[src2[i]];
|
|
}
|
|
src += s->uvlinesize;
|
|
src2 += s->uvlinesize;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Chroma MC always uses qpel bilinear */
|
|
uvmx = (uvmx&3)<<1;
|
|
uvmy = (uvmy&3)<<1;
|
|
if(!v->rnd){
|
|
dsp->put_h264_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
dsp->put_h264_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}else{
|
|
v->vc1dsp.put_no_rnd_vc1_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
v->vc1dsp.put_no_rnd_vc1_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}
|
|
}
|
|
|
|
/***********************************************************************/
|
|
/**
|
|
* @defgroup vc1block VC-1 Block-level functions
|
|
* @see 7.1.4, p91 and 8.1.1.7, p(1)04
|
|
* @{
|
|
*/
|
|
|
|
/**
|
|
* @def GET_MQUANT
|
|
* @brief Get macroblock-level quantizer scale
|
|
*/
|
|
#define GET_MQUANT() \
|
|
if (v->dquantfrm) \
|
|
{ \
|
|
int edges = 0; \
|
|
if (v->dqprofile == DQPROFILE_ALL_MBS) \
|
|
{ \
|
|
if (v->dqbilevel) \
|
|
{ \
|
|
mquant = (get_bits1(gb)) ? v->altpq : v->pq; \
|
|
} \
|
|
else \
|
|
{ \
|
|
mqdiff = get_bits(gb, 3); \
|
|
if (mqdiff != 7) mquant = v->pq + mqdiff; \
|
|
else mquant = get_bits(gb, 5); \
|
|
} \
|
|
} \
|
|
if(v->dqprofile == DQPROFILE_SINGLE_EDGE) \
|
|
edges = 1 << v->dqsbedge; \
|
|
else if(v->dqprofile == DQPROFILE_DOUBLE_EDGES) \
|
|
edges = (3 << v->dqsbedge) % 15; \
|
|
else if(v->dqprofile == DQPROFILE_FOUR_EDGES) \
|
|
edges = 15; \
|
|
if((edges&1) && !s->mb_x) \
|
|
mquant = v->altpq; \
|
|
if((edges&2) && s->first_slice_line) \
|
|
mquant = v->altpq; \
|
|
if((edges&4) && s->mb_x == (s->mb_width - 1)) \
|
|
mquant = v->altpq; \
|
|
if((edges&8) && s->mb_y == (s->mb_height - 1)) \
|
|
mquant = v->altpq; \
|
|
}
|
|
|
|
/**
|
|
* @def GET_MVDATA(_dmv_x, _dmv_y)
|
|
* @brief Get MV differentials
|
|
* @see MVDATA decoding from 8.3.5.2, p(1)20
|
|
* @param _dmv_x Horizontal differential for decoded MV
|
|
* @param _dmv_y Vertical differential for decoded MV
|
|
*/
|
|
#define GET_MVDATA(_dmv_x, _dmv_y) \
|
|
index = 1 + get_vlc2(gb, ff_vc1_mv_diff_vlc[s->mv_table_index].table,\
|
|
VC1_MV_DIFF_VLC_BITS, 2); \
|
|
if (index > 36) \
|
|
{ \
|
|
mb_has_coeffs = 1; \
|
|
index -= 37; \
|
|
} \
|
|
else mb_has_coeffs = 0; \
|
|
s->mb_intra = 0; \
|
|
if (!index) { _dmv_x = _dmv_y = 0; } \
|
|
else if (index == 35) \
|
|
{ \
|
|
_dmv_x = get_bits(gb, v->k_x - 1 + s->quarter_sample); \
|
|
_dmv_y = get_bits(gb, v->k_y - 1 + s->quarter_sample); \
|
|
} \
|
|
else if (index == 36) \
|
|
{ \
|
|
_dmv_x = 0; \
|
|
_dmv_y = 0; \
|
|
s->mb_intra = 1; \
|
|
} \
|
|
else \
|
|
{ \
|
|
index1 = index%6; \
|
|
if (!s->quarter_sample && index1 == 5) val = 1; \
|
|
else val = 0; \
|
|
if(size_table[index1] - val > 0) \
|
|
val = get_bits(gb, size_table[index1] - val); \
|
|
else val = 0; \
|
|
sign = 0 - (val&1); \
|
|
_dmv_x = (sign ^ ((val>>1) + offset_table[index1])) - sign; \
|
|
\
|
|
index1 = index/6; \
|
|
if (!s->quarter_sample && index1 == 5) val = 1; \
|
|
else val = 0; \
|
|
if(size_table[index1] - val > 0) \
|
|
val = get_bits(gb, size_table[index1] - val); \
|
|
else val = 0; \
|
|
sign = 0 - (val&1); \
|
|
_dmv_y = (sign ^ ((val>>1) + offset_table[index1])) - sign; \
|
|
}
|
|
|
|
/** Predict and set motion vector
|
|
*/
|
|
static inline void vc1_pred_mv(VC1Context *v, int n, int dmv_x, int dmv_y, int mv1, int r_x, int r_y, uint8_t* is_intra)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int xy, wrap, off = 0;
|
|
int16_t *A, *B, *C;
|
|
int px, py;
|
|
int sum;
|
|
|
|
/* scale MV difference to be quad-pel */
|
|
dmv_x <<= 1 - s->quarter_sample;
|
|
dmv_y <<= 1 - s->quarter_sample;
|
|
|
|
wrap = s->b8_stride;
|
|
xy = s->block_index[n];
|
|
|
|
if(s->mb_intra){
|
|
s->mv[0][n][0] = s->current_picture.motion_val[0][xy][0] = 0;
|
|
s->mv[0][n][1] = s->current_picture.motion_val[0][xy][1] = 0;
|
|
s->current_picture.motion_val[1][xy][0] = 0;
|
|
s->current_picture.motion_val[1][xy][1] = 0;
|
|
if(mv1) { /* duplicate motion data for 1-MV block */
|
|
s->current_picture.motion_val[0][xy + 1][0] = 0;
|
|
s->current_picture.motion_val[0][xy + 1][1] = 0;
|
|
s->current_picture.motion_val[0][xy + wrap][0] = 0;
|
|
s->current_picture.motion_val[0][xy + wrap][1] = 0;
|
|
s->current_picture.motion_val[0][xy + wrap + 1][0] = 0;
|
|
s->current_picture.motion_val[0][xy + wrap + 1][1] = 0;
|
|
v->luma_mv[s->mb_x][0] = v->luma_mv[s->mb_x][1] = 0;
|
|
s->current_picture.motion_val[1][xy + 1][0] = 0;
|
|
s->current_picture.motion_val[1][xy + 1][1] = 0;
|
|
s->current_picture.motion_val[1][xy + wrap][0] = 0;
|
|
s->current_picture.motion_val[1][xy + wrap][1] = 0;
|
|
s->current_picture.motion_val[1][xy + wrap + 1][0] = 0;
|
|
s->current_picture.motion_val[1][xy + wrap + 1][1] = 0;
|
|
}
|
|
return;
|
|
}
|
|
|
|
C = s->current_picture.motion_val[0][xy - 1];
|
|
A = s->current_picture.motion_val[0][xy - wrap];
|
|
if(mv1)
|
|
off = (s->mb_x == (s->mb_width - 1)) ? -1 : 2;
|
|
else {
|
|
//in 4-MV mode different blocks have different B predictor position
|
|
switch(n){
|
|
case 0:
|
|
off = (s->mb_x > 0) ? -1 : 1;
|
|
break;
|
|
case 1:
|
|
off = (s->mb_x == (s->mb_width - 1)) ? -1 : 1;
|
|
break;
|
|
case 2:
|
|
off = 1;
|
|
break;
|
|
case 3:
|
|
off = -1;
|
|
}
|
|
}
|
|
B = s->current_picture.motion_val[0][xy - wrap + off];
|
|
|
|
if(!s->first_slice_line || (n==2 || n==3)) { // predictor A is not out of bounds
|
|
if(s->mb_width == 1) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = mid_pred(A[0], B[0], C[0]);
|
|
py = mid_pred(A[1], B[1], C[1]);
|
|
}
|
|
} else if(s->mb_x || (n==1 || n==3)) { // predictor C is not out of bounds
|
|
px = C[0];
|
|
py = C[1];
|
|
} else {
|
|
px = py = 0;
|
|
}
|
|
/* Pullback MV as specified in 8.3.5.3.4 */
|
|
{
|
|
int qx, qy, X, Y;
|
|
qx = (s->mb_x << 6) + ((n==1 || n==3) ? 32 : 0);
|
|
qy = (s->mb_y << 6) + ((n==2 || n==3) ? 32 : 0);
|
|
X = (s->mb_width << 6) - 4;
|
|
Y = (s->mb_height << 6) - 4;
|
|
if(mv1) {
|
|
if(qx + px < -60) px = -60 - qx;
|
|
if(qy + py < -60) py = -60 - qy;
|
|
} else {
|
|
if(qx + px < -28) px = -28 - qx;
|
|
if(qy + py < -28) py = -28 - qy;
|
|
}
|
|
if(qx + px > X) px = X - qx;
|
|
if(qy + py > Y) py = Y - qy;
|
|
}
|
|
/* Calculate hybrid prediction as specified in 8.3.5.3.5 */
|
|
if((!s->first_slice_line || (n==2 || n==3)) && (s->mb_x || (n==1 || n==3))) {
|
|
if(is_intra[xy - wrap])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - A[0]) + FFABS(py - A[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
} else {
|
|
if(is_intra[xy - 1])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - C[0]) + FFABS(py - C[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
/* store MV using signed modulus of MV range defined in 4.11 */
|
|
s->mv[0][n][0] = s->current_picture.motion_val[0][xy][0] = ((px + dmv_x + r_x) & ((r_x << 1) - 1)) - r_x;
|
|
s->mv[0][n][1] = s->current_picture.motion_val[0][xy][1] = ((py + dmv_y + r_y) & ((r_y << 1) - 1)) - r_y;
|
|
if(mv1) { /* duplicate motion data for 1-MV block */
|
|
s->current_picture.motion_val[0][xy + 1][0] = s->current_picture.motion_val[0][xy][0];
|
|
s->current_picture.motion_val[0][xy + 1][1] = s->current_picture.motion_val[0][xy][1];
|
|
s->current_picture.motion_val[0][xy + wrap][0] = s->current_picture.motion_val[0][xy][0];
|
|
s->current_picture.motion_val[0][xy + wrap][1] = s->current_picture.motion_val[0][xy][1];
|
|
s->current_picture.motion_val[0][xy + wrap + 1][0] = s->current_picture.motion_val[0][xy][0];
|
|
s->current_picture.motion_val[0][xy + wrap + 1][1] = s->current_picture.motion_val[0][xy][1];
|
|
}
|
|
}
|
|
|
|
/** Motion compensation for direct or interpolated blocks in B-frames
|
|
*/
|
|
static void vc1_interp_mc(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
DSPContext *dsp = &v->s.dsp;
|
|
uint8_t *srcY, *srcU, *srcV;
|
|
int dxy, mx, my, uvmx, uvmy, src_x, src_y, uvsrc_x, uvsrc_y;
|
|
|
|
if(!v->s.next_picture.data[0])return;
|
|
|
|
mx = s->mv[1][0][0];
|
|
my = s->mv[1][0][1];
|
|
uvmx = (mx + ((mx & 3) == 3)) >> 1;
|
|
uvmy = (my + ((my & 3) == 3)) >> 1;
|
|
if(v->fastuvmc) {
|
|
uvmx = uvmx + ((uvmx<0)?-(uvmx&1):(uvmx&1));
|
|
uvmy = uvmy + ((uvmy<0)?-(uvmy&1):(uvmy&1));
|
|
}
|
|
srcY = s->next_picture.data[0];
|
|
srcU = s->next_picture.data[1];
|
|
srcV = s->next_picture.data[2];
|
|
|
|
src_x = s->mb_x * 16 + (mx >> 2);
|
|
src_y = s->mb_y * 16 + (my >> 2);
|
|
uvsrc_x = s->mb_x * 8 + (uvmx >> 2);
|
|
uvsrc_y = s->mb_y * 8 + (uvmy >> 2);
|
|
|
|
if(v->profile != PROFILE_ADVANCED){
|
|
src_x = av_clip( src_x, -16, s->mb_width * 16);
|
|
src_y = av_clip( src_y, -16, s->mb_height * 16);
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->mb_width * 8);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->mb_height * 8);
|
|
}else{
|
|
src_x = av_clip( src_x, -17, s->avctx->coded_width);
|
|
src_y = av_clip( src_y, -18, s->avctx->coded_height + 1);
|
|
uvsrc_x = av_clip(uvsrc_x, -8, s->avctx->coded_width >> 1);
|
|
uvsrc_y = av_clip(uvsrc_y, -8, s->avctx->coded_height >> 1);
|
|
}
|
|
|
|
srcY += src_y * s->linesize + src_x;
|
|
srcU += uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
srcV += uvsrc_y * s->uvlinesize + uvsrc_x;
|
|
|
|
/* for grayscale we should not try to read from unknown area */
|
|
if(s->flags & CODEC_FLAG_GRAY) {
|
|
srcU = s->edge_emu_buffer + 18 * s->linesize;
|
|
srcV = s->edge_emu_buffer + 18 * s->linesize;
|
|
}
|
|
|
|
if(v->rangeredfrm
|
|
|| (unsigned)(src_x - s->mspel) > s->h_edge_pos - (mx&3) - 16 - s->mspel*3
|
|
|| (unsigned)(src_y - s->mspel) > s->v_edge_pos - (my&3) - 16 - s->mspel*3){
|
|
uint8_t *uvbuf= s->edge_emu_buffer + 19 * s->linesize;
|
|
|
|
srcY -= s->mspel * (1 + s->linesize);
|
|
s->dsp.emulated_edge_mc(s->edge_emu_buffer, srcY, s->linesize, 17+s->mspel*2, 17+s->mspel*2,
|
|
src_x - s->mspel, src_y - s->mspel, s->h_edge_pos, s->v_edge_pos);
|
|
srcY = s->edge_emu_buffer;
|
|
s->dsp.emulated_edge_mc(uvbuf , srcU, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
s->dsp.emulated_edge_mc(uvbuf + 16, srcV, s->uvlinesize, 8+1, 8+1,
|
|
uvsrc_x, uvsrc_y, s->h_edge_pos >> 1, s->v_edge_pos >> 1);
|
|
srcU = uvbuf;
|
|
srcV = uvbuf + 16;
|
|
/* if we deal with range reduction we need to scale source blocks */
|
|
if(v->rangeredfrm) {
|
|
int i, j;
|
|
uint8_t *src, *src2;
|
|
|
|
src = srcY;
|
|
for(j = 0; j < 17 + s->mspel*2; j++) {
|
|
for(i = 0; i < 17 + s->mspel*2; i++) src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src += s->linesize;
|
|
}
|
|
src = srcU; src2 = srcV;
|
|
for(j = 0; j < 9; j++) {
|
|
for(i = 0; i < 9; i++) {
|
|
src[i] = ((src[i] - 128) >> 1) + 128;
|
|
src2[i] = ((src2[i] - 128) >> 1) + 128;
|
|
}
|
|
src += s->uvlinesize;
|
|
src2 += s->uvlinesize;
|
|
}
|
|
}
|
|
srcY += s->mspel * (1 + s->linesize);
|
|
}
|
|
|
|
if(s->mspel) {
|
|
dxy = ((my & 3) << 2) | (mx & 3);
|
|
v->vc1dsp.avg_vc1_mspel_pixels_tab[dxy](s->dest[0] , srcY , s->linesize, v->rnd);
|
|
v->vc1dsp.avg_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8, srcY + 8, s->linesize, v->rnd);
|
|
srcY += s->linesize * 8;
|
|
v->vc1dsp.avg_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8 * s->linesize , srcY , s->linesize, v->rnd);
|
|
v->vc1dsp.avg_vc1_mspel_pixels_tab[dxy](s->dest[0] + 8 * s->linesize + 8, srcY + 8, s->linesize, v->rnd);
|
|
} else { // hpel mc
|
|
dxy = (my & 2) | ((mx & 2) >> 1);
|
|
|
|
if(!v->rnd)
|
|
dsp->avg_pixels_tab[0][dxy](s->dest[0], srcY, s->linesize, 16);
|
|
else
|
|
dsp->avg_no_rnd_pixels_tab[0][dxy](s->dest[0], srcY, s->linesize, 16);
|
|
}
|
|
|
|
if(s->flags & CODEC_FLAG_GRAY) return;
|
|
/* Chroma MC always uses qpel blilinear */
|
|
uvmx = (uvmx&3)<<1;
|
|
uvmy = (uvmy&3)<<1;
|
|
if(!v->rnd){
|
|
dsp->avg_h264_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
dsp->avg_h264_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}else{
|
|
v->vc1dsp.avg_no_rnd_vc1_chroma_pixels_tab[0](s->dest[1], srcU, s->uvlinesize, 8, uvmx, uvmy);
|
|
v->vc1dsp.avg_no_rnd_vc1_chroma_pixels_tab[0](s->dest[2], srcV, s->uvlinesize, 8, uvmx, uvmy);
|
|
}
|
|
}
|
|
|
|
static av_always_inline int scale_mv(int value, int bfrac, int inv, int qs)
|
|
{
|
|
int n = bfrac;
|
|
|
|
#if B_FRACTION_DEN==256
|
|
if(inv)
|
|
n -= 256;
|
|
if(!qs)
|
|
return 2 * ((value * n + 255) >> 9);
|
|
return (value * n + 128) >> 8;
|
|
#else
|
|
if(inv)
|
|
n -= B_FRACTION_DEN;
|
|
if(!qs)
|
|
return 2 * ((value * n + B_FRACTION_DEN - 1) / (2 * B_FRACTION_DEN));
|
|
return (value * n + B_FRACTION_DEN/2) / B_FRACTION_DEN;
|
|
#endif
|
|
}
|
|
|
|
/** Reconstruct motion vector for B-frame and do motion compensation
|
|
*/
|
|
static inline void vc1_b_mc(VC1Context *v, int dmv_x[2], int dmv_y[2], int direct, int mode)
|
|
{
|
|
if(v->use_ic) {
|
|
v->mv_mode2 = v->mv_mode;
|
|
v->mv_mode = MV_PMODE_INTENSITY_COMP;
|
|
}
|
|
if(direct) {
|
|
vc1_mc_1mv(v, 0);
|
|
vc1_interp_mc(v);
|
|
if(v->use_ic) v->mv_mode = v->mv_mode2;
|
|
return;
|
|
}
|
|
if(mode == BMV_TYPE_INTERPOLATED) {
|
|
vc1_mc_1mv(v, 0);
|
|
vc1_interp_mc(v);
|
|
if(v->use_ic) v->mv_mode = v->mv_mode2;
|
|
return;
|
|
}
|
|
|
|
if(v->use_ic && (mode == BMV_TYPE_BACKWARD)) v->mv_mode = v->mv_mode2;
|
|
vc1_mc_1mv(v, (mode == BMV_TYPE_BACKWARD));
|
|
if(v->use_ic) v->mv_mode = v->mv_mode2;
|
|
}
|
|
|
|
static inline void vc1_pred_b_mv(VC1Context *v, int dmv_x[2], int dmv_y[2], int direct, int mvtype)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int xy, wrap, off = 0;
|
|
int16_t *A, *B, *C;
|
|
int px, py;
|
|
int sum;
|
|
int r_x, r_y;
|
|
const uint8_t *is_intra = v->mb_type[0];
|
|
|
|
r_x = v->range_x;
|
|
r_y = v->range_y;
|
|
/* scale MV difference to be quad-pel */
|
|
dmv_x[0] <<= 1 - s->quarter_sample;
|
|
dmv_y[0] <<= 1 - s->quarter_sample;
|
|
dmv_x[1] <<= 1 - s->quarter_sample;
|
|
dmv_y[1] <<= 1 - s->quarter_sample;
|
|
|
|
wrap = s->b8_stride;
|
|
xy = s->block_index[0];
|
|
|
|
if(s->mb_intra) {
|
|
s->current_picture.motion_val[0][xy][0] =
|
|
s->current_picture.motion_val[0][xy][1] =
|
|
s->current_picture.motion_val[1][xy][0] =
|
|
s->current_picture.motion_val[1][xy][1] = 0;
|
|
return;
|
|
}
|
|
s->mv[0][0][0] = scale_mv(s->next_picture.motion_val[1][xy][0], v->bfraction, 0, s->quarter_sample);
|
|
s->mv[0][0][1] = scale_mv(s->next_picture.motion_val[1][xy][1], v->bfraction, 0, s->quarter_sample);
|
|
s->mv[1][0][0] = scale_mv(s->next_picture.motion_val[1][xy][0], v->bfraction, 1, s->quarter_sample);
|
|
s->mv[1][0][1] = scale_mv(s->next_picture.motion_val[1][xy][1], v->bfraction, 1, s->quarter_sample);
|
|
|
|
/* Pullback predicted motion vectors as specified in 8.4.5.4 */
|
|
s->mv[0][0][0] = av_clip(s->mv[0][0][0], -60 - (s->mb_x << 6), (s->mb_width << 6) - 4 - (s->mb_x << 6));
|
|
s->mv[0][0][1] = av_clip(s->mv[0][0][1], -60 - (s->mb_y << 6), (s->mb_height << 6) - 4 - (s->mb_y << 6));
|
|
s->mv[1][0][0] = av_clip(s->mv[1][0][0], -60 - (s->mb_x << 6), (s->mb_width << 6) - 4 - (s->mb_x << 6));
|
|
s->mv[1][0][1] = av_clip(s->mv[1][0][1], -60 - (s->mb_y << 6), (s->mb_height << 6) - 4 - (s->mb_y << 6));
|
|
if(direct) {
|
|
s->current_picture.motion_val[0][xy][0] = s->mv[0][0][0];
|
|
s->current_picture.motion_val[0][xy][1] = s->mv[0][0][1];
|
|
s->current_picture.motion_val[1][xy][0] = s->mv[1][0][0];
|
|
s->current_picture.motion_val[1][xy][1] = s->mv[1][0][1];
|
|
return;
|
|
}
|
|
|
|
if((mvtype == BMV_TYPE_FORWARD) || (mvtype == BMV_TYPE_INTERPOLATED)) {
|
|
C = s->current_picture.motion_val[0][xy - 2];
|
|
A = s->current_picture.motion_val[0][xy - wrap*2];
|
|
off = (s->mb_x == (s->mb_width - 1)) ? -2 : 2;
|
|
B = s->current_picture.motion_val[0][xy - wrap*2 + off];
|
|
|
|
if(!s->mb_x) C[0] = C[1] = 0;
|
|
if(!s->first_slice_line) { // predictor A is not out of bounds
|
|
if(s->mb_width == 1) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = mid_pred(A[0], B[0], C[0]);
|
|
py = mid_pred(A[1], B[1], C[1]);
|
|
}
|
|
} else if(s->mb_x) { // predictor C is not out of bounds
|
|
px = C[0];
|
|
py = C[1];
|
|
} else {
|
|
px = py = 0;
|
|
}
|
|
/* Pullback MV as specified in 8.3.5.3.4 */
|
|
{
|
|
int qx, qy, X, Y;
|
|
if(v->profile < PROFILE_ADVANCED) {
|
|
qx = (s->mb_x << 5);
|
|
qy = (s->mb_y << 5);
|
|
X = (s->mb_width << 5) - 4;
|
|
Y = (s->mb_height << 5) - 4;
|
|
if(qx + px < -28) px = -28 - qx;
|
|
if(qy + py < -28) py = -28 - qy;
|
|
if(qx + px > X) px = X - qx;
|
|
if(qy + py > Y) py = Y - qy;
|
|
} else {
|
|
qx = (s->mb_x << 6);
|
|
qy = (s->mb_y << 6);
|
|
X = (s->mb_width << 6) - 4;
|
|
Y = (s->mb_height << 6) - 4;
|
|
if(qx + px < -60) px = -60 - qx;
|
|
if(qy + py < -60) py = -60 - qy;
|
|
if(qx + px > X) px = X - qx;
|
|
if(qy + py > Y) py = Y - qy;
|
|
}
|
|
}
|
|
/* Calculate hybrid prediction as specified in 8.3.5.3.5 */
|
|
if(0 && !s->first_slice_line && s->mb_x) {
|
|
if(is_intra[xy - wrap])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - A[0]) + FFABS(py - A[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
} else {
|
|
if(is_intra[xy - 2])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - C[0]) + FFABS(py - C[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
/* store MV using signed modulus of MV range defined in 4.11 */
|
|
s->mv[0][0][0] = ((px + dmv_x[0] + r_x) & ((r_x << 1) - 1)) - r_x;
|
|
s->mv[0][0][1] = ((py + dmv_y[0] + r_y) & ((r_y << 1) - 1)) - r_y;
|
|
}
|
|
if((mvtype == BMV_TYPE_BACKWARD) || (mvtype == BMV_TYPE_INTERPOLATED)) {
|
|
C = s->current_picture.motion_val[1][xy - 2];
|
|
A = s->current_picture.motion_val[1][xy - wrap*2];
|
|
off = (s->mb_x == (s->mb_width - 1)) ? -2 : 2;
|
|
B = s->current_picture.motion_val[1][xy - wrap*2 + off];
|
|
|
|
if(!s->mb_x) C[0] = C[1] = 0;
|
|
if(!s->first_slice_line) { // predictor A is not out of bounds
|
|
if(s->mb_width == 1) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = mid_pred(A[0], B[0], C[0]);
|
|
py = mid_pred(A[1], B[1], C[1]);
|
|
}
|
|
} else if(s->mb_x) { // predictor C is not out of bounds
|
|
px = C[0];
|
|
py = C[1];
|
|
} else {
|
|
px = py = 0;
|
|
}
|
|
/* Pullback MV as specified in 8.3.5.3.4 */
|
|
{
|
|
int qx, qy, X, Y;
|
|
if(v->profile < PROFILE_ADVANCED) {
|
|
qx = (s->mb_x << 5);
|
|
qy = (s->mb_y << 5);
|
|
X = (s->mb_width << 5) - 4;
|
|
Y = (s->mb_height << 5) - 4;
|
|
if(qx + px < -28) px = -28 - qx;
|
|
if(qy + py < -28) py = -28 - qy;
|
|
if(qx + px > X) px = X - qx;
|
|
if(qy + py > Y) py = Y - qy;
|
|
} else {
|
|
qx = (s->mb_x << 6);
|
|
qy = (s->mb_y << 6);
|
|
X = (s->mb_width << 6) - 4;
|
|
Y = (s->mb_height << 6) - 4;
|
|
if(qx + px < -60) px = -60 - qx;
|
|
if(qy + py < -60) py = -60 - qy;
|
|
if(qx + px > X) px = X - qx;
|
|
if(qy + py > Y) py = Y - qy;
|
|
}
|
|
}
|
|
/* Calculate hybrid prediction as specified in 8.3.5.3.5 */
|
|
if(0 && !s->first_slice_line && s->mb_x) {
|
|
if(is_intra[xy - wrap])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - A[0]) + FFABS(py - A[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
} else {
|
|
if(is_intra[xy - 2])
|
|
sum = FFABS(px) + FFABS(py);
|
|
else
|
|
sum = FFABS(px - C[0]) + FFABS(py - C[1]);
|
|
if(sum > 32) {
|
|
if(get_bits1(&s->gb)) {
|
|
px = A[0];
|
|
py = A[1];
|
|
} else {
|
|
px = C[0];
|
|
py = C[1];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
/* store MV using signed modulus of MV range defined in 4.11 */
|
|
|
|
s->mv[1][0][0] = ((px + dmv_x[1] + r_x) & ((r_x << 1) - 1)) - r_x;
|
|
s->mv[1][0][1] = ((py + dmv_y[1] + r_y) & ((r_y << 1) - 1)) - r_y;
|
|
}
|
|
s->current_picture.motion_val[0][xy][0] = s->mv[0][0][0];
|
|
s->current_picture.motion_val[0][xy][1] = s->mv[0][0][1];
|
|
s->current_picture.motion_val[1][xy][0] = s->mv[1][0][0];
|
|
s->current_picture.motion_val[1][xy][1] = s->mv[1][0][1];
|
|
}
|
|
|
|
/** Get predicted DC value for I-frames only
|
|
* prediction dir: left=0, top=1
|
|
* @param s MpegEncContext
|
|
* @param overlap flag indicating that overlap filtering is used
|
|
* @param pq integer part of picture quantizer
|
|
* @param[in] n block index in the current MB
|
|
* @param dc_val_ptr Pointer to DC predictor
|
|
* @param dir_ptr Prediction direction for use in AC prediction
|
|
*/
|
|
static inline int vc1_i_pred_dc(MpegEncContext *s, int overlap, int pq, int n,
|
|
int16_t **dc_val_ptr, int *dir_ptr)
|
|
{
|
|
int a, b, c, wrap, pred, scale;
|
|
int16_t *dc_val;
|
|
static const uint16_t dcpred[32] = {
|
|
-1, 1024, 512, 341, 256, 205, 171, 146, 128,
|
|
114, 102, 93, 85, 79, 73, 68, 64,
|
|
60, 57, 54, 51, 49, 47, 45, 43,
|
|
41, 39, 38, 37, 35, 34, 33
|
|
};
|
|
|
|
/* find prediction - wmv3_dc_scale always used here in fact */
|
|
if (n < 4) scale = s->y_dc_scale;
|
|
else scale = s->c_dc_scale;
|
|
|
|
wrap = s->block_wrap[n];
|
|
dc_val= s->dc_val[0] + s->block_index[n];
|
|
|
|
/* B A
|
|
* C X
|
|
*/
|
|
c = dc_val[ - 1];
|
|
b = dc_val[ - 1 - wrap];
|
|
a = dc_val[ - wrap];
|
|
|
|
if (pq < 9 || !overlap)
|
|
{
|
|
/* Set outer values */
|
|
if (s->first_slice_line && (n!=2 && n!=3)) b=a=dcpred[scale];
|
|
if (s->mb_x == 0 && (n!=1 && n!=3)) b=c=dcpred[scale];
|
|
}
|
|
else
|
|
{
|
|
/* Set outer values */
|
|
if (s->first_slice_line && (n!=2 && n!=3)) b=a=0;
|
|
if (s->mb_x == 0 && (n!=1 && n!=3)) b=c=0;
|
|
}
|
|
|
|
if (abs(a - b) <= abs(b - c)) {
|
|
pred = c;
|
|
*dir_ptr = 1;//left
|
|
} else {
|
|
pred = a;
|
|
*dir_ptr = 0;//top
|
|
}
|
|
|
|
/* update predictor */
|
|
*dc_val_ptr = &dc_val[0];
|
|
return pred;
|
|
}
|
|
|
|
|
|
/** Get predicted DC value
|
|
* prediction dir: left=0, top=1
|
|
* @param s MpegEncContext
|
|
* @param overlap flag indicating that overlap filtering is used
|
|
* @param pq integer part of picture quantizer
|
|
* @param[in] n block index in the current MB
|
|
* @param a_avail flag indicating top block availability
|
|
* @param c_avail flag indicating left block availability
|
|
* @param dc_val_ptr Pointer to DC predictor
|
|
* @param dir_ptr Prediction direction for use in AC prediction
|
|
*/
|
|
static inline int vc1_pred_dc(MpegEncContext *s, int overlap, int pq, int n,
|
|
int a_avail, int c_avail,
|
|
int16_t **dc_val_ptr, int *dir_ptr)
|
|
{
|
|
int a, b, c, wrap, pred;
|
|
int16_t *dc_val;
|
|
int mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
int q1, q2 = 0;
|
|
|
|
wrap = s->block_wrap[n];
|
|
dc_val= s->dc_val[0] + s->block_index[n];
|
|
|
|
/* B A
|
|
* C X
|
|
*/
|
|
c = dc_val[ - 1];
|
|
b = dc_val[ - 1 - wrap];
|
|
a = dc_val[ - wrap];
|
|
/* scale predictors if needed */
|
|
q1 = s->current_picture.qscale_table[mb_pos];
|
|
if(c_avail && (n!= 1 && n!=3)) {
|
|
q2 = s->current_picture.qscale_table[mb_pos - 1];
|
|
if(q2 && q2 != q1)
|
|
c = (c * s->y_dc_scale_table[q2] * ff_vc1_dqscale[s->y_dc_scale_table[q1] - 1] + 0x20000) >> 18;
|
|
}
|
|
if(a_avail && (n!= 2 && n!=3)) {
|
|
q2 = s->current_picture.qscale_table[mb_pos - s->mb_stride];
|
|
if(q2 && q2 != q1)
|
|
a = (a * s->y_dc_scale_table[q2] * ff_vc1_dqscale[s->y_dc_scale_table[q1] - 1] + 0x20000) >> 18;
|
|
}
|
|
if(a_avail && c_avail && (n!=3)) {
|
|
int off = mb_pos;
|
|
if(n != 1) off--;
|
|
if(n != 2) off -= s->mb_stride;
|
|
q2 = s->current_picture.qscale_table[off];
|
|
if(q2 && q2 != q1)
|
|
b = (b * s->y_dc_scale_table[q2] * ff_vc1_dqscale[s->y_dc_scale_table[q1] - 1] + 0x20000) >> 18;
|
|
}
|
|
|
|
if(a_avail && c_avail) {
|
|
if(abs(a - b) <= abs(b - c)) {
|
|
pred = c;
|
|
*dir_ptr = 1;//left
|
|
} else {
|
|
pred = a;
|
|
*dir_ptr = 0;//top
|
|
}
|
|
} else if(a_avail) {
|
|
pred = a;
|
|
*dir_ptr = 0;//top
|
|
} else if(c_avail) {
|
|
pred = c;
|
|
*dir_ptr = 1;//left
|
|
} else {
|
|
pred = 0;
|
|
*dir_ptr = 1;//left
|
|
}
|
|
|
|
/* update predictor */
|
|
*dc_val_ptr = &dc_val[0];
|
|
return pred;
|
|
}
|
|
|
|
/** @} */ // Block group
|
|
|
|
/**
|
|
* @defgroup vc1_std_mb VC1 Macroblock-level functions in Simple/Main Profiles
|
|
* @see 7.1.4, p91 and 8.1.1.7, p(1)04
|
|
* @{
|
|
*/
|
|
|
|
static inline int vc1_coded_block_pred(MpegEncContext * s, int n, uint8_t **coded_block_ptr)
|
|
{
|
|
int xy, wrap, pred, a, b, c;
|
|
|
|
xy = s->block_index[n];
|
|
wrap = s->b8_stride;
|
|
|
|
/* B C
|
|
* A X
|
|
*/
|
|
a = s->coded_block[xy - 1 ];
|
|
b = s->coded_block[xy - 1 - wrap];
|
|
c = s->coded_block[xy - wrap];
|
|
|
|
if (b == c) {
|
|
pred = a;
|
|
} else {
|
|
pred = c;
|
|
}
|
|
|
|
/* store value */
|
|
*coded_block_ptr = &s->coded_block[xy];
|
|
|
|
return pred;
|
|
}
|
|
|
|
/**
|
|
* Decode one AC coefficient
|
|
* @param v The VC1 context
|
|
* @param last Last coefficient
|
|
* @param skip How much zero coefficients to skip
|
|
* @param value Decoded AC coefficient value
|
|
* @param codingset set of VLC to decode data
|
|
* @see 8.1.3.4
|
|
*/
|
|
static void vc1_decode_ac_coeff(VC1Context *v, int *last, int *skip, int *value, int codingset)
|
|
{
|
|
GetBitContext *gb = &v->s.gb;
|
|
int index, escape, run = 0, level = 0, lst = 0;
|
|
|
|
index = get_vlc2(gb, ff_vc1_ac_coeff_table[codingset].table, AC_VLC_BITS, 3);
|
|
if (index != vc1_ac_sizes[codingset] - 1) {
|
|
run = vc1_index_decode_table[codingset][index][0];
|
|
level = vc1_index_decode_table[codingset][index][1];
|
|
lst = index >= vc1_last_decode_table[codingset] || get_bits_left(gb) < 0;
|
|
if(get_bits1(gb))
|
|
level = -level;
|
|
} else {
|
|
escape = decode210(gb);
|
|
if (escape != 2) {
|
|
index = get_vlc2(gb, ff_vc1_ac_coeff_table[codingset].table, AC_VLC_BITS, 3);
|
|
run = vc1_index_decode_table[codingset][index][0];
|
|
level = vc1_index_decode_table[codingset][index][1];
|
|
lst = index >= vc1_last_decode_table[codingset];
|
|
if(escape == 0) {
|
|
if(lst)
|
|
level += vc1_last_delta_level_table[codingset][run];
|
|
else
|
|
level += vc1_delta_level_table[codingset][run];
|
|
} else {
|
|
if(lst)
|
|
run += vc1_last_delta_run_table[codingset][level] + 1;
|
|
else
|
|
run += vc1_delta_run_table[codingset][level] + 1;
|
|
}
|
|
if(get_bits1(gb))
|
|
level = -level;
|
|
} else {
|
|
int sign;
|
|
lst = get_bits1(gb);
|
|
if(v->s.esc3_level_length == 0) {
|
|
if(v->pq < 8 || v->dquantfrm) { // table 59
|
|
v->s.esc3_level_length = get_bits(gb, 3);
|
|
if(!v->s.esc3_level_length)
|
|
v->s.esc3_level_length = get_bits(gb, 2) + 8;
|
|
} else { //table 60
|
|
v->s.esc3_level_length = get_unary(gb, 1, 6) + 2;
|
|
}
|
|
v->s.esc3_run_length = 3 + get_bits(gb, 2);
|
|
}
|
|
run = get_bits(gb, v->s.esc3_run_length);
|
|
sign = get_bits1(gb);
|
|
level = get_bits(gb, v->s.esc3_level_length);
|
|
if(sign)
|
|
level = -level;
|
|
}
|
|
}
|
|
|
|
*last = lst;
|
|
*skip = run;
|
|
*value = level;
|
|
}
|
|
|
|
/** Decode intra block in intra frames - should be faster than decode_intra_block
|
|
* @param v VC1Context
|
|
* @param block block to decode
|
|
* @param[in] n subblock index
|
|
* @param coded are AC coeffs present or not
|
|
* @param codingset set of VLC to decode data
|
|
*/
|
|
static int vc1_decode_i_block(VC1Context *v, DCTELEM block[64], int n, int coded, int codingset)
|
|
{
|
|
GetBitContext *gb = &v->s.gb;
|
|
MpegEncContext *s = &v->s;
|
|
int dc_pred_dir = 0; /* Direction of the DC prediction used */
|
|
int i;
|
|
int16_t *dc_val;
|
|
int16_t *ac_val, *ac_val2;
|
|
int dcdiff;
|
|
|
|
/* Get DC differential */
|
|
if (n < 4) {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_luma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
} else {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_chroma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
}
|
|
if (dcdiff < 0){
|
|
av_log(s->avctx, AV_LOG_ERROR, "Illegal DC VLC\n");
|
|
return -1;
|
|
}
|
|
if (dcdiff)
|
|
{
|
|
if (dcdiff == 119 /* ESC index value */)
|
|
{
|
|
/* TODO: Optimize */
|
|
if (v->pq == 1) dcdiff = get_bits(gb, 10);
|
|
else if (v->pq == 2) dcdiff = get_bits(gb, 9);
|
|
else dcdiff = get_bits(gb, 8);
|
|
}
|
|
else
|
|
{
|
|
if (v->pq == 1)
|
|
dcdiff = (dcdiff<<2) + get_bits(gb, 2) - 3;
|
|
else if (v->pq == 2)
|
|
dcdiff = (dcdiff<<1) + get_bits1(gb) - 1;
|
|
}
|
|
if (get_bits1(gb))
|
|
dcdiff = -dcdiff;
|
|
}
|
|
|
|
/* Prediction */
|
|
dcdiff += vc1_i_pred_dc(&v->s, v->overlap, v->pq, n, &dc_val, &dc_pred_dir);
|
|
*dc_val = dcdiff;
|
|
|
|
/* Store the quantized DC coeff, used for prediction */
|
|
if (n < 4) {
|
|
block[0] = dcdiff * s->y_dc_scale;
|
|
} else {
|
|
block[0] = dcdiff * s->c_dc_scale;
|
|
}
|
|
/* Skip ? */
|
|
if (!coded) {
|
|
goto not_coded;
|
|
}
|
|
|
|
//AC Decoding
|
|
i = 1;
|
|
|
|
{
|
|
int last = 0, skip, value;
|
|
const uint8_t *zz_table;
|
|
int scale;
|
|
int k;
|
|
|
|
scale = v->pq * 2 + v->halfpq;
|
|
|
|
if(v->s.ac_pred) {
|
|
if(!dc_pred_dir)
|
|
zz_table = v->zz_8x8[2];
|
|
else
|
|
zz_table = v->zz_8x8[3];
|
|
} else
|
|
zz_table = v->zz_8x8[1];
|
|
|
|
ac_val = s->ac_val[0][0] + s->block_index[n] * 16;
|
|
ac_val2 = ac_val;
|
|
if(dc_pred_dir) //left
|
|
ac_val -= 16;
|
|
else //top
|
|
ac_val -= 16 * s->block_wrap[n];
|
|
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, codingset);
|
|
i += skip;
|
|
if(i > 63)
|
|
break;
|
|
block[zz_table[i++]] = value;
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(s->ac_pred) {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->left_blk_sh] += ac_val[k];
|
|
} else { //top
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->top_blk_sh] += ac_val[k + 8];
|
|
}
|
|
}
|
|
/* save AC coeffs for further prediction */
|
|
for(k = 1; k < 8; k++) {
|
|
ac_val2[k] = block[k << v->left_blk_sh];
|
|
ac_val2[k + 8] = block[k << v->top_blk_sh];
|
|
}
|
|
|
|
/* scale AC coeffs */
|
|
for(k = 1; k < 64; k++)
|
|
if(block[k]) {
|
|
block[k] *= scale;
|
|
if(!v->pquantizer)
|
|
block[k] += (block[k] < 0) ? -v->pq : v->pq;
|
|
}
|
|
|
|
if(s->ac_pred) i = 63;
|
|
}
|
|
|
|
not_coded:
|
|
if(!coded) {
|
|
int k, scale;
|
|
ac_val = s->ac_val[0][0] + s->block_index[n] * 16;
|
|
ac_val2 = ac_val;
|
|
|
|
i = 0;
|
|
scale = v->pq * 2 + v->halfpq;
|
|
memset(ac_val2, 0, 16 * 2);
|
|
if(dc_pred_dir) {//left
|
|
ac_val -= 16;
|
|
if(s->ac_pred)
|
|
memcpy(ac_val2, ac_val, 8 * 2);
|
|
} else {//top
|
|
ac_val -= 16 * s->block_wrap[n];
|
|
if(s->ac_pred)
|
|
memcpy(ac_val2 + 8, ac_val + 8, 8 * 2);
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(s->ac_pred) {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->left_blk_sh] = ac_val[k] * scale;
|
|
if(!v->pquantizer && block[k << v->left_blk_sh])
|
|
block[k << v->left_blk_sh] += (block[k << v->left_blk_sh] < 0) ? -v->pq : v->pq;
|
|
}
|
|
} else { //top
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->top_blk_sh] = ac_val[k + 8] * scale;
|
|
if(!v->pquantizer && block[k << v->top_blk_sh])
|
|
block[k << v->top_blk_sh] += (block[k << v->top_blk_sh] < 0) ? -v->pq : v->pq;
|
|
}
|
|
}
|
|
i = 63;
|
|
}
|
|
}
|
|
s->block_last_index[n] = i;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/** Decode intra block in intra frames - should be faster than decode_intra_block
|
|
* @param v VC1Context
|
|
* @param block block to decode
|
|
* @param[in] n subblock number
|
|
* @param coded are AC coeffs present or not
|
|
* @param codingset set of VLC to decode data
|
|
* @param mquant quantizer value for this macroblock
|
|
*/
|
|
static int vc1_decode_i_block_adv(VC1Context *v, DCTELEM block[64], int n, int coded, int codingset, int mquant)
|
|
{
|
|
GetBitContext *gb = &v->s.gb;
|
|
MpegEncContext *s = &v->s;
|
|
int dc_pred_dir = 0; /* Direction of the DC prediction used */
|
|
int i;
|
|
int16_t *dc_val;
|
|
int16_t *ac_val, *ac_val2;
|
|
int dcdiff;
|
|
int a_avail = v->a_avail, c_avail = v->c_avail;
|
|
int use_pred = s->ac_pred;
|
|
int scale;
|
|
int q1, q2 = 0;
|
|
int mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
|
|
/* Get DC differential */
|
|
if (n < 4) {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_luma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
} else {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_chroma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
}
|
|
if (dcdiff < 0){
|
|
av_log(s->avctx, AV_LOG_ERROR, "Illegal DC VLC\n");
|
|
return -1;
|
|
}
|
|
if (dcdiff)
|
|
{
|
|
if (dcdiff == 119 /* ESC index value */)
|
|
{
|
|
/* TODO: Optimize */
|
|
if (mquant == 1) dcdiff = get_bits(gb, 10);
|
|
else if (mquant == 2) dcdiff = get_bits(gb, 9);
|
|
else dcdiff = get_bits(gb, 8);
|
|
}
|
|
else
|
|
{
|
|
if (mquant == 1)
|
|
dcdiff = (dcdiff<<2) + get_bits(gb, 2) - 3;
|
|
else if (mquant == 2)
|
|
dcdiff = (dcdiff<<1) + get_bits1(gb) - 1;
|
|
}
|
|
if (get_bits1(gb))
|
|
dcdiff = -dcdiff;
|
|
}
|
|
|
|
/* Prediction */
|
|
dcdiff += vc1_pred_dc(&v->s, v->overlap, mquant, n, v->a_avail, v->c_avail, &dc_val, &dc_pred_dir);
|
|
*dc_val = dcdiff;
|
|
|
|
/* Store the quantized DC coeff, used for prediction */
|
|
if (n < 4) {
|
|
block[0] = dcdiff * s->y_dc_scale;
|
|
} else {
|
|
block[0] = dcdiff * s->c_dc_scale;
|
|
}
|
|
|
|
//AC Decoding
|
|
i = 1;
|
|
|
|
/* check if AC is needed at all */
|
|
if(!a_avail && !c_avail) use_pred = 0;
|
|
ac_val = s->ac_val[0][0] + s->block_index[n] * 16;
|
|
ac_val2 = ac_val;
|
|
|
|
scale = mquant * 2 + ((mquant == v->pq) ? v->halfpq : 0);
|
|
|
|
if(dc_pred_dir) //left
|
|
ac_val -= 16;
|
|
else //top
|
|
ac_val -= 16 * s->block_wrap[n];
|
|
|
|
q1 = s->current_picture.qscale_table[mb_pos];
|
|
if(dc_pred_dir && c_avail && mb_pos) q2 = s->current_picture.qscale_table[mb_pos - 1];
|
|
if(!dc_pred_dir && a_avail && mb_pos >= s->mb_stride) q2 = s->current_picture.qscale_table[mb_pos - s->mb_stride];
|
|
if(dc_pred_dir && n==1) q2 = q1;
|
|
if(!dc_pred_dir && n==2) q2 = q1;
|
|
if(n==3) q2 = q1;
|
|
|
|
if(coded) {
|
|
int last = 0, skip, value;
|
|
const uint8_t *zz_table;
|
|
int k;
|
|
|
|
if(v->s.ac_pred) {
|
|
if(!dc_pred_dir)
|
|
zz_table = v->zz_8x8[2];
|
|
else
|
|
zz_table = v->zz_8x8[3];
|
|
} else
|
|
zz_table = v->zz_8x8[1];
|
|
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, codingset);
|
|
i += skip;
|
|
if(i > 63)
|
|
break;
|
|
block[zz_table[i++]] = value;
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(use_pred) {
|
|
/* scale predictors if needed*/
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->left_blk_sh] += (ac_val[k] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
} else { //top
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->top_blk_sh] += (ac_val[k + 8] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
} else {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->left_blk_sh] += ac_val[k];
|
|
} else { //top
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->top_blk_sh] += ac_val[k + 8];
|
|
}
|
|
}
|
|
}
|
|
/* save AC coeffs for further prediction */
|
|
for(k = 1; k < 8; k++) {
|
|
ac_val2[k ] = block[k << v->left_blk_sh];
|
|
ac_val2[k + 8] = block[k << v->top_blk_sh];
|
|
}
|
|
|
|
/* scale AC coeffs */
|
|
for(k = 1; k < 64; k++)
|
|
if(block[k]) {
|
|
block[k] *= scale;
|
|
if(!v->pquantizer)
|
|
block[k] += (block[k] < 0) ? -mquant : mquant;
|
|
}
|
|
|
|
if(use_pred) i = 63;
|
|
} else { // no AC coeffs
|
|
int k;
|
|
|
|
memset(ac_val2, 0, 16 * 2);
|
|
if(dc_pred_dir) {//left
|
|
if(use_pred) {
|
|
memcpy(ac_val2, ac_val, 8 * 2);
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
for(k = 1; k < 8; k++)
|
|
ac_val2[k] = (ac_val2[k] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
}
|
|
} else {//top
|
|
if(use_pred) {
|
|
memcpy(ac_val2 + 8, ac_val + 8, 8 * 2);
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
for(k = 1; k < 8; k++)
|
|
ac_val2[k + 8] = (ac_val2[k + 8] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(use_pred) {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->left_blk_sh] = ac_val2[k] * scale;
|
|
if(!v->pquantizer && block[k << v->left_blk_sh])
|
|
block[k << v->left_blk_sh] += (block[k << v->left_blk_sh] < 0) ? -mquant : mquant;
|
|
}
|
|
} else { //top
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->top_blk_sh] = ac_val2[k + 8] * scale;
|
|
if(!v->pquantizer && block[k << v->top_blk_sh])
|
|
block[k << v->top_blk_sh] += (block[k << v->top_blk_sh] < 0) ? -mquant : mquant;
|
|
}
|
|
}
|
|
i = 63;
|
|
}
|
|
}
|
|
s->block_last_index[n] = i;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/** Decode intra block in inter frames - more generic version than vc1_decode_i_block
|
|
* @param v VC1Context
|
|
* @param block block to decode
|
|
* @param[in] n subblock index
|
|
* @param coded are AC coeffs present or not
|
|
* @param mquant block quantizer
|
|
* @param codingset set of VLC to decode data
|
|
*/
|
|
static int vc1_decode_intra_block(VC1Context *v, DCTELEM block[64], int n, int coded, int mquant, int codingset)
|
|
{
|
|
GetBitContext *gb = &v->s.gb;
|
|
MpegEncContext *s = &v->s;
|
|
int dc_pred_dir = 0; /* Direction of the DC prediction used */
|
|
int i;
|
|
int16_t *dc_val;
|
|
int16_t *ac_val, *ac_val2;
|
|
int dcdiff;
|
|
int mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
int a_avail = v->a_avail, c_avail = v->c_avail;
|
|
int use_pred = s->ac_pred;
|
|
int scale;
|
|
int q1, q2 = 0;
|
|
|
|
s->dsp.clear_block(block);
|
|
|
|
/* XXX: Guard against dumb values of mquant */
|
|
mquant = (mquant < 1) ? 0 : ( (mquant>31) ? 31 : mquant );
|
|
|
|
/* Set DC scale - y and c use the same */
|
|
s->y_dc_scale = s->y_dc_scale_table[mquant];
|
|
s->c_dc_scale = s->c_dc_scale_table[mquant];
|
|
|
|
/* Get DC differential */
|
|
if (n < 4) {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_luma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
} else {
|
|
dcdiff = get_vlc2(&s->gb, ff_msmp4_dc_chroma_vlc[s->dc_table_index].table, DC_VLC_BITS, 3);
|
|
}
|
|
if (dcdiff < 0){
|
|
av_log(s->avctx, AV_LOG_ERROR, "Illegal DC VLC\n");
|
|
return -1;
|
|
}
|
|
if (dcdiff)
|
|
{
|
|
if (dcdiff == 119 /* ESC index value */)
|
|
{
|
|
/* TODO: Optimize */
|
|
if (mquant == 1) dcdiff = get_bits(gb, 10);
|
|
else if (mquant == 2) dcdiff = get_bits(gb, 9);
|
|
else dcdiff = get_bits(gb, 8);
|
|
}
|
|
else
|
|
{
|
|
if (mquant == 1)
|
|
dcdiff = (dcdiff<<2) + get_bits(gb, 2) - 3;
|
|
else if (mquant == 2)
|
|
dcdiff = (dcdiff<<1) + get_bits1(gb) - 1;
|
|
}
|
|
if (get_bits1(gb))
|
|
dcdiff = -dcdiff;
|
|
}
|
|
|
|
/* Prediction */
|
|
dcdiff += vc1_pred_dc(&v->s, v->overlap, mquant, n, a_avail, c_avail, &dc_val, &dc_pred_dir);
|
|
*dc_val = dcdiff;
|
|
|
|
/* Store the quantized DC coeff, used for prediction */
|
|
|
|
if (n < 4) {
|
|
block[0] = dcdiff * s->y_dc_scale;
|
|
} else {
|
|
block[0] = dcdiff * s->c_dc_scale;
|
|
}
|
|
|
|
//AC Decoding
|
|
i = 1;
|
|
|
|
/* check if AC is needed at all and adjust direction if needed */
|
|
if(!a_avail) dc_pred_dir = 1;
|
|
if(!c_avail) dc_pred_dir = 0;
|
|
if(!a_avail && !c_avail) use_pred = 0;
|
|
ac_val = s->ac_val[0][0] + s->block_index[n] * 16;
|
|
ac_val2 = ac_val;
|
|
|
|
scale = mquant * 2 + v->halfpq;
|
|
|
|
if(dc_pred_dir) //left
|
|
ac_val -= 16;
|
|
else //top
|
|
ac_val -= 16 * s->block_wrap[n];
|
|
|
|
q1 = s->current_picture.qscale_table[mb_pos];
|
|
if(dc_pred_dir && c_avail && mb_pos) q2 = s->current_picture.qscale_table[mb_pos - 1];
|
|
if(!dc_pred_dir && a_avail && mb_pos >= s->mb_stride) q2 = s->current_picture.qscale_table[mb_pos - s->mb_stride];
|
|
if(dc_pred_dir && n==1) q2 = q1;
|
|
if(!dc_pred_dir && n==2) q2 = q1;
|
|
if(n==3) q2 = q1;
|
|
|
|
if(coded) {
|
|
int last = 0, skip, value;
|
|
int k;
|
|
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, codingset);
|
|
i += skip;
|
|
if(i > 63)
|
|
break;
|
|
block[v->zz_8x8[0][i++]] = value;
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(use_pred) {
|
|
/* scale predictors if needed*/
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->left_blk_sh] += (ac_val[k] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
} else { //top
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->top_blk_sh] += (ac_val[k + 8] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
} else {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->left_blk_sh] += ac_val[k];
|
|
} else { //top
|
|
for(k = 1; k < 8; k++)
|
|
block[k << v->top_blk_sh] += ac_val[k + 8];
|
|
}
|
|
}
|
|
}
|
|
/* save AC coeffs for further prediction */
|
|
for(k = 1; k < 8; k++) {
|
|
ac_val2[k ] = block[k << v->left_blk_sh];
|
|
ac_val2[k + 8] = block[k << v->top_blk_sh];
|
|
}
|
|
|
|
/* scale AC coeffs */
|
|
for(k = 1; k < 64; k++)
|
|
if(block[k]) {
|
|
block[k] *= scale;
|
|
if(!v->pquantizer)
|
|
block[k] += (block[k] < 0) ? -mquant : mquant;
|
|
}
|
|
|
|
if(use_pred) i = 63;
|
|
} else { // no AC coeffs
|
|
int k;
|
|
|
|
memset(ac_val2, 0, 16 * 2);
|
|
if(dc_pred_dir) {//left
|
|
if(use_pred) {
|
|
memcpy(ac_val2, ac_val, 8 * 2);
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
for(k = 1; k < 8; k++)
|
|
ac_val2[k] = (ac_val2[k] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
}
|
|
} else {//top
|
|
if(use_pred) {
|
|
memcpy(ac_val2 + 8, ac_val + 8, 8 * 2);
|
|
if(q2 && q1!=q2) {
|
|
q1 = q1 * 2 + ((q1 == v->pq) ? v->halfpq : 0) - 1;
|
|
q2 = q2 * 2 + ((q2 == v->pq) ? v->halfpq : 0) - 1;
|
|
for(k = 1; k < 8; k++)
|
|
ac_val2[k + 8] = (ac_val2[k + 8] * q2 * ff_vc1_dqscale[q1 - 1] + 0x20000) >> 18;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* apply AC prediction if needed */
|
|
if(use_pred) {
|
|
if(dc_pred_dir) { //left
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->left_blk_sh] = ac_val2[k] * scale;
|
|
if(!v->pquantizer && block[k << v->left_blk_sh])
|
|
block[k << v->left_blk_sh] += (block[k << v->left_blk_sh] < 0) ? -mquant : mquant;
|
|
}
|
|
} else { //top
|
|
for(k = 1; k < 8; k++) {
|
|
block[k << v->top_blk_sh] = ac_val2[k + 8] * scale;
|
|
if(!v->pquantizer && block[k << v->top_blk_sh])
|
|
block[k << v->top_blk_sh] += (block[k << v->top_blk_sh] < 0) ? -mquant : mquant;
|
|
}
|
|
}
|
|
i = 63;
|
|
}
|
|
}
|
|
s->block_last_index[n] = i;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/** Decode P block
|
|
*/
|
|
static int vc1_decode_p_block(VC1Context *v, DCTELEM block[64], int n, int mquant, int ttmb, int first_block,
|
|
uint8_t *dst, int linesize, int skip_block, int *ttmb_out)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
GetBitContext *gb = &s->gb;
|
|
int i, j;
|
|
int subblkpat = 0;
|
|
int scale, off, idx, last, skip, value;
|
|
int ttblk = ttmb & 7;
|
|
int pat = 0;
|
|
|
|
s->dsp.clear_block(block);
|
|
|
|
if(ttmb == -1) {
|
|
ttblk = ff_vc1_ttblk_to_tt[v->tt_index][get_vlc2(gb, ff_vc1_ttblk_vlc[v->tt_index].table, VC1_TTBLK_VLC_BITS, 1)];
|
|
}
|
|
if(ttblk == TT_4X4) {
|
|
subblkpat = ~(get_vlc2(gb, ff_vc1_subblkpat_vlc[v->tt_index].table, VC1_SUBBLKPAT_VLC_BITS, 1) + 1);
|
|
}
|
|
if((ttblk != TT_8X8 && ttblk != TT_4X4)
|
|
&& ((v->ttmbf || (ttmb != -1 && (ttmb & 8) && !first_block))
|
|
|| (!v->res_rtm_flag && !first_block))) {
|
|
subblkpat = decode012(gb);
|
|
if(subblkpat) subblkpat ^= 3; //swap decoded pattern bits
|
|
if(ttblk == TT_8X4_TOP || ttblk == TT_8X4_BOTTOM) ttblk = TT_8X4;
|
|
if(ttblk == TT_4X8_RIGHT || ttblk == TT_4X8_LEFT) ttblk = TT_4X8;
|
|
}
|
|
scale = 2 * mquant + ((v->pq == mquant) ? v->halfpq : 0);
|
|
|
|
// convert transforms like 8X4_TOP to generic TT and SUBBLKPAT
|
|
if(ttblk == TT_8X4_TOP || ttblk == TT_8X4_BOTTOM) {
|
|
subblkpat = 2 - (ttblk == TT_8X4_TOP);
|
|
ttblk = TT_8X4;
|
|
}
|
|
if(ttblk == TT_4X8_RIGHT || ttblk == TT_4X8_LEFT) {
|
|
subblkpat = 2 - (ttblk == TT_4X8_LEFT);
|
|
ttblk = TT_4X8;
|
|
}
|
|
switch(ttblk) {
|
|
case TT_8X8:
|
|
pat = 0xF;
|
|
i = 0;
|
|
last = 0;
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, v->codingset2);
|
|
i += skip;
|
|
if(i > 63)
|
|
break;
|
|
idx = v->zz_8x8[0][i++];
|
|
block[idx] = value * scale;
|
|
if(!v->pquantizer)
|
|
block[idx] += (block[idx] < 0) ? -mquant : mquant;
|
|
}
|
|
if(!skip_block){
|
|
if(i==1)
|
|
v->vc1dsp.vc1_inv_trans_8x8_dc(dst, linesize, block);
|
|
else{
|
|
v->vc1dsp.vc1_inv_trans_8x8(block);
|
|
s->dsp.add_pixels_clamped(block, dst, linesize);
|
|
}
|
|
}
|
|
break;
|
|
case TT_4X4:
|
|
pat = ~subblkpat & 0xF;
|
|
for(j = 0; j < 4; j++) {
|
|
last = subblkpat & (1 << (3 - j));
|
|
i = 0;
|
|
off = (j & 1) * 4 + (j & 2) * 16;
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, v->codingset2);
|
|
i += skip;
|
|
if(i > 15)
|
|
break;
|
|
idx = ff_vc1_simple_progressive_4x4_zz[i++];
|
|
block[idx + off] = value * scale;
|
|
if(!v->pquantizer)
|
|
block[idx + off] += (block[idx + off] < 0) ? -mquant : mquant;
|
|
}
|
|
if(!(subblkpat & (1 << (3 - j))) && !skip_block){
|
|
if(i==1)
|
|
v->vc1dsp.vc1_inv_trans_4x4_dc(dst + (j&1)*4 + (j&2)*2*linesize, linesize, block + off);
|
|
else
|
|
v->vc1dsp.vc1_inv_trans_4x4(dst + (j&1)*4 + (j&2)*2*linesize, linesize, block + off);
|
|
}
|
|
}
|
|
break;
|
|
case TT_8X4:
|
|
pat = ~((subblkpat & 2)*6 + (subblkpat & 1)*3) & 0xF;
|
|
for(j = 0; j < 2; j++) {
|
|
last = subblkpat & (1 << (1 - j));
|
|
i = 0;
|
|
off = j * 32;
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, v->codingset2);
|
|
i += skip;
|
|
if(i > 31)
|
|
break;
|
|
idx = v->zz_8x4[i++]+off;
|
|
block[idx] = value * scale;
|
|
if(!v->pquantizer)
|
|
block[idx] += (block[idx] < 0) ? -mquant : mquant;
|
|
}
|
|
if(!(subblkpat & (1 << (1 - j))) && !skip_block){
|
|
if(i==1)
|
|
v->vc1dsp.vc1_inv_trans_8x4_dc(dst + j*4*linesize, linesize, block + off);
|
|
else
|
|
v->vc1dsp.vc1_inv_trans_8x4(dst + j*4*linesize, linesize, block + off);
|
|
}
|
|
}
|
|
break;
|
|
case TT_4X8:
|
|
pat = ~(subblkpat*5) & 0xF;
|
|
for(j = 0; j < 2; j++) {
|
|
last = subblkpat & (1 << (1 - j));
|
|
i = 0;
|
|
off = j * 4;
|
|
while (!last) {
|
|
vc1_decode_ac_coeff(v, &last, &skip, &value, v->codingset2);
|
|
i += skip;
|
|
if(i > 31)
|
|
break;
|
|
idx = v->zz_4x8[i++]+off;
|
|
block[idx] = value * scale;
|
|
if(!v->pquantizer)
|
|
block[idx] += (block[idx] < 0) ? -mquant : mquant;
|
|
}
|
|
if(!(subblkpat & (1 << (1 - j))) && !skip_block){
|
|
if(i==1)
|
|
v->vc1dsp.vc1_inv_trans_4x8_dc(dst + j*4, linesize, block + off);
|
|
else
|
|
v->vc1dsp.vc1_inv_trans_4x8(dst + j*4, linesize, block + off);
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
if (ttmb_out)
|
|
*ttmb_out |= ttblk << (n * 4);
|
|
return pat;
|
|
}
|
|
|
|
/** @} */ // Macroblock group
|
|
|
|
static const int size_table [6] = { 0, 2, 3, 4, 5, 8 };
|
|
static const int offset_table[6] = { 0, 1, 3, 7, 15, 31 };
|
|
|
|
static av_always_inline void vc1_apply_p_v_loop_filter(VC1Context *v, int block_num)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int mb_cbp = v->cbp[s->mb_x - s->mb_stride],
|
|
block_cbp = mb_cbp >> (block_num * 4), bottom_cbp,
|
|
mb_is_intra = v->is_intra[s->mb_x - s->mb_stride],
|
|
block_is_intra = mb_is_intra >> (block_num * 4), bottom_is_intra;
|
|
int idx, linesize = block_num > 3 ? s->uvlinesize : s->linesize, ttblk;
|
|
uint8_t *dst;
|
|
|
|
if(block_num > 3) {
|
|
dst = s->dest[block_num - 3];
|
|
} else {
|
|
dst = s->dest[0] + (block_num & 1) * 8 + ((block_num & 2) * 4 - 8) * linesize;
|
|
}
|
|
if (s->mb_y != s->mb_height || block_num < 2) {
|
|
int16_t (*mv)[2];
|
|
int mv_stride;
|
|
|
|
if(block_num > 3) {
|
|
bottom_cbp = v->cbp[s->mb_x] >> (block_num * 4);
|
|
bottom_is_intra = v->is_intra[s->mb_x] >> (block_num * 4);
|
|
mv = &v->luma_mv[s->mb_x - s->mb_stride];
|
|
mv_stride = s->mb_stride;
|
|
} else {
|
|
bottom_cbp = (block_num < 2) ? (mb_cbp >> ((block_num + 2) * 4)) :
|
|
(v->cbp[s->mb_x] >> ((block_num - 2) * 4));
|
|
bottom_is_intra = (block_num < 2) ? (mb_is_intra >> ((block_num + 2) * 4)) :
|
|
(v->is_intra[s->mb_x] >> ((block_num - 2) * 4));
|
|
mv_stride = s->b8_stride;
|
|
mv = &s->current_picture.motion_val[0][s->block_index[block_num] - 2 * mv_stride];
|
|
}
|
|
|
|
if (bottom_is_intra & 1 || block_is_intra & 1 ||
|
|
mv[0][0] != mv[mv_stride][0] || mv[0][1] != mv[mv_stride][1]) {
|
|
v->vc1dsp.vc1_v_loop_filter8(dst, linesize, v->pq);
|
|
} else {
|
|
idx = ((bottom_cbp >> 2) | block_cbp) & 3;
|
|
if(idx == 3) {
|
|
v->vc1dsp.vc1_v_loop_filter8(dst, linesize, v->pq);
|
|
} else if (idx) {
|
|
if (idx == 1)
|
|
v->vc1dsp.vc1_v_loop_filter4(dst + 4, linesize, v->pq);
|
|
else
|
|
v->vc1dsp.vc1_v_loop_filter4(dst, linesize, v->pq);
|
|
}
|
|
}
|
|
}
|
|
|
|
dst -= 4 * linesize;
|
|
ttblk = (v->ttblk[s->mb_x - s->mb_stride] >> (block_num * 4)) & 0xf;
|
|
if (ttblk == TT_4X4 || ttblk == TT_8X4) {
|
|
idx = (block_cbp | (block_cbp >> 2)) & 3;
|
|
if (idx == 3) {
|
|
v->vc1dsp.vc1_v_loop_filter8(dst, linesize, v->pq);
|
|
} else if (idx) {
|
|
if (idx == 1)
|
|
v->vc1dsp.vc1_v_loop_filter4(dst + 4, linesize, v->pq);
|
|
else
|
|
v->vc1dsp.vc1_v_loop_filter4(dst, linesize, v->pq);
|
|
}
|
|
}
|
|
}
|
|
|
|
static av_always_inline void vc1_apply_p_h_loop_filter(VC1Context *v, int block_num)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int mb_cbp = v->cbp[s->mb_x - 1 - s->mb_stride],
|
|
block_cbp = mb_cbp >> (block_num * 4), right_cbp,
|
|
mb_is_intra = v->is_intra[s->mb_x - 1 - s->mb_stride],
|
|
block_is_intra = mb_is_intra >> (block_num * 4), right_is_intra;
|
|
int idx, linesize = block_num > 3 ? s->uvlinesize : s->linesize, ttblk;
|
|
uint8_t *dst;
|
|
|
|
if (block_num > 3) {
|
|
dst = s->dest[block_num - 3] - 8 * linesize;
|
|
} else {
|
|
dst = s->dest[0] + (block_num & 1) * 8 + ((block_num & 2) * 4 - 16) * linesize - 8;
|
|
}
|
|
|
|
if (s->mb_x != s->mb_width || !(block_num & 5)) {
|
|
int16_t (*mv)[2];
|
|
|
|
if(block_num > 3) {
|
|
right_cbp = v->cbp[s->mb_x - s->mb_stride] >> (block_num * 4);
|
|
right_is_intra = v->is_intra[s->mb_x - s->mb_stride] >> (block_num * 4);
|
|
mv = &v->luma_mv[s->mb_x - s->mb_stride - 1];
|
|
}else{
|
|
right_cbp = (block_num & 1) ? (v->cbp[s->mb_x - s->mb_stride] >> ((block_num - 1) * 4)) :
|
|
(mb_cbp >> ((block_num + 1) * 4));
|
|
right_is_intra = (block_num & 1) ? (v->is_intra[s->mb_x - s->mb_stride] >> ((block_num - 1) * 4)) :
|
|
(mb_is_intra >> ((block_num + 1) * 4));
|
|
mv = &s->current_picture.motion_val[0][s->block_index[block_num] - s->b8_stride * 2 - 2];
|
|
}
|
|
if (block_is_intra & 1 || right_is_intra & 1 || mv[0][0] != mv[1][0] || mv[0][1] != mv[1][1]) {
|
|
v->vc1dsp.vc1_h_loop_filter8(dst, linesize, v->pq);
|
|
} else {
|
|
idx = ((right_cbp >> 1) | block_cbp) & 5; // FIXME check
|
|
if (idx == 5) {
|
|
v->vc1dsp.vc1_h_loop_filter8(dst, linesize, v->pq);
|
|
} else if (idx) {
|
|
if (idx == 1)
|
|
v->vc1dsp.vc1_h_loop_filter4(dst+4*linesize, linesize, v->pq);
|
|
else
|
|
v->vc1dsp.vc1_h_loop_filter4(dst, linesize, v->pq);
|
|
}
|
|
}
|
|
}
|
|
|
|
dst -= 4;
|
|
ttblk = (v->ttblk[s->mb_x - s->mb_stride - 1] >> (block_num * 4)) & 0xf;
|
|
if (ttblk == TT_4X4 || ttblk == TT_4X8) {
|
|
idx = (block_cbp | (block_cbp >> 1)) & 5;
|
|
if (idx == 5) {
|
|
v->vc1dsp.vc1_h_loop_filter8(dst, linesize, v->pq);
|
|
} else if (idx) {
|
|
if (idx == 1)
|
|
v->vc1dsp.vc1_h_loop_filter4(dst + linesize*4, linesize, v->pq);
|
|
else
|
|
v->vc1dsp.vc1_h_loop_filter4(dst, linesize, v->pq);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void vc1_apply_p_loop_filter(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int i;
|
|
|
|
for (i = 0; i < 6; i++) {
|
|
vc1_apply_p_v_loop_filter(v, i);
|
|
}
|
|
|
|
/* V always preceedes H, therefore we run H one MB before V;
|
|
* at the end of a row, we catch up to complete the row */
|
|
if (s->mb_x) {
|
|
for (i = 0; i < 6; i++) {
|
|
vc1_apply_p_h_loop_filter(v, i);
|
|
}
|
|
if (s->mb_x == s->mb_width - 1) {
|
|
s->mb_x++;
|
|
ff_update_block_index(s);
|
|
for (i = 0; i < 6; i++) {
|
|
vc1_apply_p_h_loop_filter(v, i);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Decode one P-frame MB (in Simple/Main profile)
|
|
*/
|
|
static int vc1_decode_p_mb(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
GetBitContext *gb = &s->gb;
|
|
int i, j;
|
|
int mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
int cbp; /* cbp decoding stuff */
|
|
int mqdiff, mquant; /* MB quantization */
|
|
int ttmb = v->ttfrm; /* MB Transform type */
|
|
|
|
int mb_has_coeffs = 1; /* last_flag */
|
|
int dmv_x, dmv_y; /* Differential MV components */
|
|
int index, index1; /* LUT indexes */
|
|
int val, sign; /* temp values */
|
|
int first_block = 1;
|
|
int dst_idx, off;
|
|
int skipped, fourmv;
|
|
int block_cbp = 0, pat, block_tt = 0, block_intra = 0;
|
|
|
|
mquant = v->pq; /* Loosy initialization */
|
|
|
|
if (v->mv_type_is_raw)
|
|
fourmv = get_bits1(gb);
|
|
else
|
|
fourmv = v->mv_type_mb_plane[mb_pos];
|
|
if (v->skip_is_raw)
|
|
skipped = get_bits1(gb);
|
|
else
|
|
skipped = v->s.mbskip_table[mb_pos];
|
|
|
|
if (!fourmv) /* 1MV mode */
|
|
{
|
|
if (!skipped)
|
|
{
|
|
GET_MVDATA(dmv_x, dmv_y);
|
|
|
|
if (s->mb_intra) {
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = 0;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = 0;
|
|
}
|
|
s->current_picture.mb_type[mb_pos] = s->mb_intra ? MB_TYPE_INTRA : MB_TYPE_16x16;
|
|
vc1_pred_mv(v, 0, dmv_x, dmv_y, 1, v->range_x, v->range_y, v->mb_type[0]);
|
|
|
|
/* FIXME Set DC val for inter block ? */
|
|
if (s->mb_intra && !mb_has_coeffs)
|
|
{
|
|
GET_MQUANT();
|
|
s->ac_pred = get_bits1(gb);
|
|
cbp = 0;
|
|
}
|
|
else if (mb_has_coeffs)
|
|
{
|
|
if (s->mb_intra) s->ac_pred = get_bits1(gb);
|
|
cbp = get_vlc2(&v->s.gb, v->cbpcy_vlc->table, VC1_CBPCY_P_VLC_BITS, 2);
|
|
GET_MQUANT();
|
|
}
|
|
else
|
|
{
|
|
mquant = v->pq;
|
|
cbp = 0;
|
|
}
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
|
|
if (!v->ttmbf && !s->mb_intra && mb_has_coeffs)
|
|
ttmb = get_vlc2(gb, ff_vc1_ttmb_vlc[v->tt_index].table,
|
|
VC1_TTMB_VLC_BITS, 2);
|
|
if(!s->mb_intra) vc1_mc_1mv(v, 0);
|
|
dst_idx = 0;
|
|
for (i=0; i<6; i++)
|
|
{
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
dst_idx += i >> 2;
|
|
val = ((cbp >> (5 - i)) & 1);
|
|
off = (i & 4) ? 0 : ((i & 1) * 8 + (i & 2) * 4 * s->linesize);
|
|
v->mb_type[0][s->block_index[i]] = s->mb_intra;
|
|
if(s->mb_intra) {
|
|
/* check if prediction blocks A and C are available */
|
|
v->a_avail = v->c_avail = 0;
|
|
if(i == 2 || i == 3 || !s->first_slice_line)
|
|
v->a_avail = v->mb_type[0][s->block_index[i] - s->block_wrap[i]];
|
|
if(i == 1 || i == 3 || s->mb_x)
|
|
v->c_avail = v->mb_type[0][s->block_index[i] - 1];
|
|
|
|
vc1_decode_intra_block(v, s->block[i], i, val, mquant, (i&4)?v->codingset2:v->codingset);
|
|
if((i>3) && (s->flags & CODEC_FLAG_GRAY)) continue;
|
|
v->vc1dsp.vc1_inv_trans_8x8(s->block[i]);
|
|
if(v->rangeredfrm) for(j = 0; j < 64; j++) s->block[i][j] <<= 1;
|
|
s->dsp.put_signed_pixels_clamped(s->block[i], s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
if(v->pq >= 9 && v->overlap) {
|
|
if(v->c_avail)
|
|
v->vc1dsp.vc1_h_overlap(s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
if(v->a_avail)
|
|
v->vc1dsp.vc1_v_overlap(s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
}
|
|
block_cbp |= 0xF << (i << 2);
|
|
block_intra |= 1 << i;
|
|
} else if(val) {
|
|
pat = vc1_decode_p_block(v, s->block[i], i, mquant, ttmb, first_block, s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize, (i&4) && (s->flags & CODEC_FLAG_GRAY), &block_tt);
|
|
block_cbp |= pat << (i << 2);
|
|
if(!v->ttmbf && ttmb < 8) ttmb = -1;
|
|
first_block = 0;
|
|
}
|
|
}
|
|
}
|
|
else //Skipped
|
|
{
|
|
s->mb_intra = 0;
|
|
for(i = 0; i < 6; i++) {
|
|
v->mb_type[0][s->block_index[i]] = 0;
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
}
|
|
s->current_picture.mb_type[mb_pos] = MB_TYPE_SKIP;
|
|
s->current_picture.qscale_table[mb_pos] = 0;
|
|
vc1_pred_mv(v, 0, 0, 0, 1, v->range_x, v->range_y, v->mb_type[0]);
|
|
vc1_mc_1mv(v, 0);
|
|
}
|
|
} //1MV mode
|
|
else //4MV mode
|
|
{
|
|
if (!skipped /* unskipped MB */)
|
|
{
|
|
int intra_count = 0, coded_inter = 0;
|
|
int is_intra[6], is_coded[6];
|
|
/* Get CBPCY */
|
|
cbp = get_vlc2(&v->s.gb, v->cbpcy_vlc->table, VC1_CBPCY_P_VLC_BITS, 2);
|
|
for (i=0; i<6; i++)
|
|
{
|
|
val = ((cbp >> (5 - i)) & 1);
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
s->mb_intra = 0;
|
|
if(i < 4) {
|
|
dmv_x = dmv_y = 0;
|
|
s->mb_intra = 0;
|
|
mb_has_coeffs = 0;
|
|
if(val) {
|
|
GET_MVDATA(dmv_x, dmv_y);
|
|
}
|
|
vc1_pred_mv(v, i, dmv_x, dmv_y, 0, v->range_x, v->range_y, v->mb_type[0]);
|
|
if(!s->mb_intra) vc1_mc_4mv_luma(v, i);
|
|
intra_count += s->mb_intra;
|
|
is_intra[i] = s->mb_intra;
|
|
is_coded[i] = mb_has_coeffs;
|
|
}
|
|
if(i&4){
|
|
is_intra[i] = (intra_count >= 3);
|
|
is_coded[i] = val;
|
|
}
|
|
if(i == 4) vc1_mc_4mv_chroma(v);
|
|
v->mb_type[0][s->block_index[i]] = is_intra[i];
|
|
if(!coded_inter) coded_inter = !is_intra[i] & is_coded[i];
|
|
}
|
|
// if there are no coded blocks then don't do anything more
|
|
dst_idx = 0;
|
|
if(!intra_count && !coded_inter)
|
|
goto end;
|
|
GET_MQUANT();
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
/* test if block is intra and has pred */
|
|
{
|
|
int intrapred = 0;
|
|
for(i=0; i<6; i++)
|
|
if(is_intra[i]) {
|
|
if(((!s->first_slice_line || (i==2 || i==3)) && v->mb_type[0][s->block_index[i] - s->block_wrap[i]])
|
|
|| ((s->mb_x || (i==1 || i==3)) && v->mb_type[0][s->block_index[i] - 1])) {
|
|
intrapred = 1;
|
|
break;
|
|
}
|
|
}
|
|
if(intrapred)s->ac_pred = get_bits1(gb);
|
|
else s->ac_pred = 0;
|
|
}
|
|
if (!v->ttmbf && coded_inter)
|
|
ttmb = get_vlc2(gb, ff_vc1_ttmb_vlc[v->tt_index].table, VC1_TTMB_VLC_BITS, 2);
|
|
for (i=0; i<6; i++)
|
|
{
|
|
dst_idx += i >> 2;
|
|
off = (i & 4) ? 0 : ((i & 1) * 8 + (i & 2) * 4 * s->linesize);
|
|
s->mb_intra = is_intra[i];
|
|
if (is_intra[i]) {
|
|
/* check if prediction blocks A and C are available */
|
|
v->a_avail = v->c_avail = 0;
|
|
if(i == 2 || i == 3 || !s->first_slice_line)
|
|
v->a_avail = v->mb_type[0][s->block_index[i] - s->block_wrap[i]];
|
|
if(i == 1 || i == 3 || s->mb_x)
|
|
v->c_avail = v->mb_type[0][s->block_index[i] - 1];
|
|
|
|
vc1_decode_intra_block(v, s->block[i], i, is_coded[i], mquant, (i&4)?v->codingset2:v->codingset);
|
|
if((i>3) && (s->flags & CODEC_FLAG_GRAY)) continue;
|
|
v->vc1dsp.vc1_inv_trans_8x8(s->block[i]);
|
|
if(v->rangeredfrm) for(j = 0; j < 64; j++) s->block[i][j] <<= 1;
|
|
s->dsp.put_signed_pixels_clamped(s->block[i], s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize);
|
|
if(v->pq >= 9 && v->overlap) {
|
|
if(v->c_avail)
|
|
v->vc1dsp.vc1_h_overlap(s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
if(v->a_avail)
|
|
v->vc1dsp.vc1_v_overlap(s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
}
|
|
block_cbp |= 0xF << (i << 2);
|
|
block_intra |= 1 << i;
|
|
} else if(is_coded[i]) {
|
|
pat = vc1_decode_p_block(v, s->block[i], i, mquant, ttmb, first_block, s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize, (i&4) && (s->flags & CODEC_FLAG_GRAY), &block_tt);
|
|
block_cbp |= pat << (i << 2);
|
|
if(!v->ttmbf && ttmb < 8) ttmb = -1;
|
|
first_block = 0;
|
|
}
|
|
}
|
|
}
|
|
else //Skipped MB
|
|
{
|
|
s->mb_intra = 0;
|
|
s->current_picture.qscale_table[mb_pos] = 0;
|
|
for (i=0; i<6; i++) {
|
|
v->mb_type[0][s->block_index[i]] = 0;
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
}
|
|
for (i=0; i<4; i++)
|
|
{
|
|
vc1_pred_mv(v, i, 0, 0, 0, v->range_x, v->range_y, v->mb_type[0]);
|
|
vc1_mc_4mv_luma(v, i);
|
|
}
|
|
vc1_mc_4mv_chroma(v);
|
|
s->current_picture.qscale_table[mb_pos] = 0;
|
|
}
|
|
}
|
|
end:
|
|
v->cbp[s->mb_x] = block_cbp;
|
|
v->ttblk[s->mb_x] = block_tt;
|
|
v->is_intra[s->mb_x] = block_intra;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/** Decode one B-frame MB (in Main profile)
|
|
*/
|
|
static void vc1_decode_b_mb(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
GetBitContext *gb = &s->gb;
|
|
int i, j;
|
|
int mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
int cbp = 0; /* cbp decoding stuff */
|
|
int mqdiff, mquant; /* MB quantization */
|
|
int ttmb = v->ttfrm; /* MB Transform type */
|
|
int mb_has_coeffs = 0; /* last_flag */
|
|
int index, index1; /* LUT indexes */
|
|
int val, sign; /* temp values */
|
|
int first_block = 1;
|
|
int dst_idx, off;
|
|
int skipped, direct;
|
|
int dmv_x[2], dmv_y[2];
|
|
int bmvtype = BMV_TYPE_BACKWARD;
|
|
|
|
mquant = v->pq; /* Loosy initialization */
|
|
s->mb_intra = 0;
|
|
|
|
if (v->dmb_is_raw)
|
|
direct = get_bits1(gb);
|
|
else
|
|
direct = v->direct_mb_plane[mb_pos];
|
|
if (v->skip_is_raw)
|
|
skipped = get_bits1(gb);
|
|
else
|
|
skipped = v->s.mbskip_table[mb_pos];
|
|
|
|
dmv_x[0] = dmv_x[1] = dmv_y[0] = dmv_y[1] = 0;
|
|
for(i = 0; i < 6; i++) {
|
|
v->mb_type[0][s->block_index[i]] = 0;
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
}
|
|
s->current_picture.qscale_table[mb_pos] = 0;
|
|
|
|
if (!direct) {
|
|
if (!skipped) {
|
|
GET_MVDATA(dmv_x[0], dmv_y[0]);
|
|
dmv_x[1] = dmv_x[0];
|
|
dmv_y[1] = dmv_y[0];
|
|
}
|
|
if(skipped || !s->mb_intra) {
|
|
bmvtype = decode012(gb);
|
|
switch(bmvtype) {
|
|
case 0:
|
|
bmvtype = (v->bfraction >= (B_FRACTION_DEN/2)) ? BMV_TYPE_BACKWARD : BMV_TYPE_FORWARD;
|
|
break;
|
|
case 1:
|
|
bmvtype = (v->bfraction >= (B_FRACTION_DEN/2)) ? BMV_TYPE_FORWARD : BMV_TYPE_BACKWARD;
|
|
break;
|
|
case 2:
|
|
bmvtype = BMV_TYPE_INTERPOLATED;
|
|
dmv_x[0] = dmv_y[0] = 0;
|
|
}
|
|
}
|
|
}
|
|
for(i = 0; i < 6; i++)
|
|
v->mb_type[0][s->block_index[i]] = s->mb_intra;
|
|
|
|
if (skipped) {
|
|
if(direct) bmvtype = BMV_TYPE_INTERPOLATED;
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
vc1_b_mc(v, dmv_x, dmv_y, direct, bmvtype);
|
|
return;
|
|
}
|
|
if (direct) {
|
|
cbp = get_vlc2(&v->s.gb, v->cbpcy_vlc->table, VC1_CBPCY_P_VLC_BITS, 2);
|
|
GET_MQUANT();
|
|
s->mb_intra = 0;
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
if(!v->ttmbf)
|
|
ttmb = get_vlc2(gb, ff_vc1_ttmb_vlc[v->tt_index].table, VC1_TTMB_VLC_BITS, 2);
|
|
dmv_x[0] = dmv_y[0] = dmv_x[1] = dmv_y[1] = 0;
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
vc1_b_mc(v, dmv_x, dmv_y, direct, bmvtype);
|
|
} else {
|
|
if(!mb_has_coeffs && !s->mb_intra) {
|
|
/* no coded blocks - effectively skipped */
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
vc1_b_mc(v, dmv_x, dmv_y, direct, bmvtype);
|
|
return;
|
|
}
|
|
if(s->mb_intra && !mb_has_coeffs) {
|
|
GET_MQUANT();
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
s->ac_pred = get_bits1(gb);
|
|
cbp = 0;
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
} else {
|
|
if(bmvtype == BMV_TYPE_INTERPOLATED) {
|
|
GET_MVDATA(dmv_x[0], dmv_y[0]);
|
|
if(!mb_has_coeffs) {
|
|
/* interpolated skipped block */
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
vc1_b_mc(v, dmv_x, dmv_y, direct, bmvtype);
|
|
return;
|
|
}
|
|
}
|
|
vc1_pred_b_mv(v, dmv_x, dmv_y, direct, bmvtype);
|
|
if(!s->mb_intra) {
|
|
vc1_b_mc(v, dmv_x, dmv_y, direct, bmvtype);
|
|
}
|
|
if(s->mb_intra)
|
|
s->ac_pred = get_bits1(gb);
|
|
cbp = get_vlc2(&v->s.gb, v->cbpcy_vlc->table, VC1_CBPCY_P_VLC_BITS, 2);
|
|
GET_MQUANT();
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
if(!v->ttmbf && !s->mb_intra && mb_has_coeffs)
|
|
ttmb = get_vlc2(gb, ff_vc1_ttmb_vlc[v->tt_index].table, VC1_TTMB_VLC_BITS, 2);
|
|
}
|
|
}
|
|
dst_idx = 0;
|
|
for (i=0; i<6; i++)
|
|
{
|
|
s->dc_val[0][s->block_index[i]] = 0;
|
|
dst_idx += i >> 2;
|
|
val = ((cbp >> (5 - i)) & 1);
|
|
off = (i & 4) ? 0 : ((i & 1) * 8 + (i & 2) * 4 * s->linesize);
|
|
v->mb_type[0][s->block_index[i]] = s->mb_intra;
|
|
if(s->mb_intra) {
|
|
/* check if prediction blocks A and C are available */
|
|
v->a_avail = v->c_avail = 0;
|
|
if(i == 2 || i == 3 || !s->first_slice_line)
|
|
v->a_avail = v->mb_type[0][s->block_index[i] - s->block_wrap[i]];
|
|
if(i == 1 || i == 3 || s->mb_x)
|
|
v->c_avail = v->mb_type[0][s->block_index[i] - 1];
|
|
|
|
vc1_decode_intra_block(v, s->block[i], i, val, mquant, (i&4)?v->codingset2:v->codingset);
|
|
if((i>3) && (s->flags & CODEC_FLAG_GRAY)) continue;
|
|
v->vc1dsp.vc1_inv_trans_8x8(s->block[i]);
|
|
if(v->rangeredfrm) for(j = 0; j < 64; j++) s->block[i][j] <<= 1;
|
|
s->dsp.put_signed_pixels_clamped(s->block[i], s->dest[dst_idx] + off, i & 4 ? s->uvlinesize : s->linesize);
|
|
} else if(val) {
|
|
vc1_decode_p_block(v, s->block[i], i, mquant, ttmb, first_block, s->dest[dst_idx] + off, (i&4)?s->uvlinesize:s->linesize, (i&4) && (s->flags & CODEC_FLAG_GRAY), NULL);
|
|
if(!v->ttmbf && ttmb < 8) ttmb = -1;
|
|
first_block = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Decode blocks of I-frame
|
|
*/
|
|
static void vc1_decode_i_blocks(VC1Context *v)
|
|
{
|
|
int k, j;
|
|
MpegEncContext *s = &v->s;
|
|
int cbp, val;
|
|
uint8_t *coded_val;
|
|
int mb_pos;
|
|
|
|
/* select codingmode used for VLC tables selection */
|
|
switch(v->y_ac_table_index){
|
|
case 0:
|
|
v->codingset = (v->pqindex <= 8) ? CS_HIGH_RATE_INTRA : CS_LOW_MOT_INTRA;
|
|
break;
|
|
case 1:
|
|
v->codingset = CS_HIGH_MOT_INTRA;
|
|
break;
|
|
case 2:
|
|
v->codingset = CS_MID_RATE_INTRA;
|
|
break;
|
|
}
|
|
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset2 = (v->pqindex <= 8) ? CS_HIGH_RATE_INTER : CS_LOW_MOT_INTER;
|
|
break;
|
|
case 1:
|
|
v->codingset2 = CS_HIGH_MOT_INTER;
|
|
break;
|
|
case 2:
|
|
v->codingset2 = CS_MID_RATE_INTER;
|
|
break;
|
|
}
|
|
|
|
/* Set DC scale - y and c use the same */
|
|
s->y_dc_scale = s->y_dc_scale_table[v->pq];
|
|
s->c_dc_scale = s->c_dc_scale_table[v->pq];
|
|
|
|
//do frame decode
|
|
s->mb_x = s->mb_y = 0;
|
|
s->mb_intra = 1;
|
|
s->first_slice_line = 1;
|
|
for(s->mb_y = 0; s->mb_y < s->mb_height; s->mb_y++) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for(; s->mb_x < s->mb_width; s->mb_x++) {
|
|
uint8_t *dst[6];
|
|
ff_update_block_index(s);
|
|
dst[0] = s->dest[0];
|
|
dst[1] = dst[0] + 8;
|
|
dst[2] = s->dest[0] + s->linesize * 8;
|
|
dst[3] = dst[2] + 8;
|
|
dst[4] = s->dest[1];
|
|
dst[5] = s->dest[2];
|
|
s->dsp.clear_blocks(s->block[0]);
|
|
mb_pos = s->mb_x + s->mb_y * s->mb_width;
|
|
s->current_picture.mb_type[mb_pos] = MB_TYPE_INTRA;
|
|
s->current_picture.qscale_table[mb_pos] = v->pq;
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = 0;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = 0;
|
|
|
|
// do actual MB decoding and displaying
|
|
cbp = get_vlc2(&v->s.gb, ff_msmp4_mb_i_vlc.table, MB_INTRA_VLC_BITS, 2);
|
|
v->s.ac_pred = get_bits1(&v->s.gb);
|
|
|
|
for(k = 0; k < 6; k++) {
|
|
val = ((cbp >> (5 - k)) & 1);
|
|
|
|
if (k < 4) {
|
|
int pred = vc1_coded_block_pred(&v->s, k, &coded_val);
|
|
val = val ^ pred;
|
|
*coded_val = val;
|
|
}
|
|
cbp |= val << (5 - k);
|
|
|
|
vc1_decode_i_block(v, s->block[k], k, val, (k<4)? v->codingset : v->codingset2);
|
|
|
|
if (k > 3 && (s->flags & CODEC_FLAG_GRAY)) continue;
|
|
v->vc1dsp.vc1_inv_trans_8x8(s->block[k]);
|
|
if(v->pq >= 9 && v->overlap) {
|
|
if (v->rangeredfrm) for(j = 0; j < 64; j++) s->block[k][j] <<= 1;
|
|
s->dsp.put_signed_pixels_clamped(s->block[k], dst[k], k & 4 ? s->uvlinesize : s->linesize);
|
|
} else {
|
|
if (v->rangeredfrm) for(j = 0; j < 64; j++) s->block[k][j] = (s->block[k][j] - 64) << 1;
|
|
s->dsp.put_pixels_clamped(s->block[k], dst[k], k & 4 ? s->uvlinesize : s->linesize);
|
|
}
|
|
}
|
|
|
|
if(v->pq >= 9 && v->overlap) {
|
|
if(s->mb_x) {
|
|
v->vc1dsp.vc1_h_overlap(s->dest[0], s->linesize);
|
|
v->vc1dsp.vc1_h_overlap(s->dest[0] + 8 * s->linesize, s->linesize);
|
|
if(!(s->flags & CODEC_FLAG_GRAY)) {
|
|
v->vc1dsp.vc1_h_overlap(s->dest[1], s->uvlinesize);
|
|
v->vc1dsp.vc1_h_overlap(s->dest[2], s->uvlinesize);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_h_overlap(s->dest[0] + 8, s->linesize);
|
|
v->vc1dsp.vc1_h_overlap(s->dest[0] + 8 * s->linesize + 8, s->linesize);
|
|
if(!s->first_slice_line) {
|
|
v->vc1dsp.vc1_v_overlap(s->dest[0], s->linesize);
|
|
v->vc1dsp.vc1_v_overlap(s->dest[0] + 8, s->linesize);
|
|
if(!(s->flags & CODEC_FLAG_GRAY)) {
|
|
v->vc1dsp.vc1_v_overlap(s->dest[1], s->uvlinesize);
|
|
v->vc1dsp.vc1_v_overlap(s->dest[2], s->uvlinesize);
|
|
}
|
|
}
|
|
v->vc1dsp.vc1_v_overlap(s->dest[0] + 8 * s->linesize, s->linesize);
|
|
v->vc1dsp.vc1_v_overlap(s->dest[0] + 8 * s->linesize + 8, s->linesize);
|
|
}
|
|
if(v->s.loop_filter) vc1_loop_filter_iblk(v, v->pq);
|
|
|
|
if(get_bits_count(&s->gb) > v->bits) {
|
|
ff_er_add_slice(s, 0, 0, s->mb_x, s->mb_y, (AC_END|DC_END|MV_END));
|
|
av_log(s->avctx, AV_LOG_ERROR, "Bits overconsumption: %i > %i\n", get_bits_count(&s->gb), v->bits);
|
|
return;
|
|
}
|
|
}
|
|
if (!v->s.loop_filter)
|
|
ff_draw_horiz_band(s, s->mb_y * 16, 16);
|
|
else if (s->mb_y)
|
|
ff_draw_horiz_band(s, (s->mb_y-1) * 16, 16);
|
|
|
|
s->first_slice_line = 0;
|
|
}
|
|
if (v->s.loop_filter)
|
|
ff_draw_horiz_band(s, (s->mb_height-1)*16, 16);
|
|
ff_er_add_slice(s, 0, 0, s->mb_width - 1, s->mb_height - 1, (AC_END|DC_END|MV_END));
|
|
}
|
|
|
|
/** Decode blocks of I-frame for advanced profile
|
|
*/
|
|
static void vc1_decode_i_blocks_adv(VC1Context *v)
|
|
{
|
|
int k;
|
|
MpegEncContext *s = &v->s;
|
|
int cbp, val;
|
|
uint8_t *coded_val;
|
|
int mb_pos;
|
|
int mquant = v->pq;
|
|
int mqdiff;
|
|
GetBitContext *gb = &s->gb;
|
|
|
|
/* select codingmode used for VLC tables selection */
|
|
switch(v->y_ac_table_index){
|
|
case 0:
|
|
v->codingset = (v->pqindex <= 8) ? CS_HIGH_RATE_INTRA : CS_LOW_MOT_INTRA;
|
|
break;
|
|
case 1:
|
|
v->codingset = CS_HIGH_MOT_INTRA;
|
|
break;
|
|
case 2:
|
|
v->codingset = CS_MID_RATE_INTRA;
|
|
break;
|
|
}
|
|
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset2 = (v->pqindex <= 8) ? CS_HIGH_RATE_INTER : CS_LOW_MOT_INTER;
|
|
break;
|
|
case 1:
|
|
v->codingset2 = CS_HIGH_MOT_INTER;
|
|
break;
|
|
case 2:
|
|
v->codingset2 = CS_MID_RATE_INTER;
|
|
break;
|
|
}
|
|
|
|
//do frame decode
|
|
s->mb_x = s->mb_y = 0;
|
|
s->mb_intra = 1;
|
|
s->first_slice_line = 1;
|
|
s->mb_y = s->start_mb_y;
|
|
if (s->start_mb_y) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
memset(&s->coded_block[s->block_index[0]-s->b8_stride], 0,
|
|
s->b8_stride * sizeof(*s->coded_block));
|
|
}
|
|
for(; s->mb_y < s->end_mb_y; s->mb_y++) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for(;s->mb_x < s->mb_width; s->mb_x++) {
|
|
DCTELEM (*block)[64] = v->block[v->cur_blk_idx];
|
|
ff_update_block_index(s);
|
|
s->dsp.clear_blocks(block[0]);
|
|
mb_pos = s->mb_x + s->mb_y * s->mb_stride;
|
|
s->current_picture.mb_type[mb_pos] = MB_TYPE_INTRA;
|
|
s->current_picture.motion_val[1][s->block_index[0]][0] = 0;
|
|
s->current_picture.motion_val[1][s->block_index[0]][1] = 0;
|
|
|
|
// do actual MB decoding and displaying
|
|
cbp = get_vlc2(&v->s.gb, ff_msmp4_mb_i_vlc.table, MB_INTRA_VLC_BITS, 2);
|
|
if(v->acpred_is_raw)
|
|
v->s.ac_pred = get_bits1(&v->s.gb);
|
|
else
|
|
v->s.ac_pred = v->acpred_plane[mb_pos];
|
|
|
|
if (v->condover == CONDOVER_SELECT && v->overflg_is_raw)
|
|
v->over_flags_plane[mb_pos] = get_bits1(&v->s.gb);
|
|
|
|
GET_MQUANT();
|
|
|
|
s->current_picture.qscale_table[mb_pos] = mquant;
|
|
/* Set DC scale - y and c use the same */
|
|
s->y_dc_scale = s->y_dc_scale_table[mquant];
|
|
s->c_dc_scale = s->c_dc_scale_table[mquant];
|
|
|
|
for(k = 0; k < 6; k++) {
|
|
val = ((cbp >> (5 - k)) & 1);
|
|
|
|
if (k < 4) {
|
|
int pred = vc1_coded_block_pred(&v->s, k, &coded_val);
|
|
val = val ^ pred;
|
|
*coded_val = val;
|
|
}
|
|
cbp |= val << (5 - k);
|
|
|
|
v->a_avail = !s->first_slice_line || (k==2 || k==3);
|
|
v->c_avail = !!s->mb_x || (k==1 || k==3);
|
|
|
|
vc1_decode_i_block_adv(v, block[k], k, val, (k<4)? v->codingset : v->codingset2, mquant);
|
|
|
|
if (k > 3 && (s->flags & CODEC_FLAG_GRAY)) continue;
|
|
v->vc1dsp.vc1_inv_trans_8x8(block[k]);
|
|
}
|
|
|
|
vc1_smooth_overlap_filter_iblk(v);
|
|
vc1_put_signed_blocks_clamped(v);
|
|
if(v->s.loop_filter) vc1_loop_filter_iblk_delayed(v, v->pq);
|
|
|
|
if(get_bits_count(&s->gb) > v->bits) {
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_x, s->mb_y, (AC_END|DC_END|MV_END));
|
|
av_log(s->avctx, AV_LOG_ERROR, "Bits overconsumption: %i > %i\n", get_bits_count(&s->gb), v->bits);
|
|
return;
|
|
}
|
|
}
|
|
if (!v->s.loop_filter)
|
|
ff_draw_horiz_band(s, s->mb_y * 16, 16);
|
|
else if (s->mb_y)
|
|
ff_draw_horiz_band(s, (s->mb_y-1) * 16, 16);
|
|
s->first_slice_line = 0;
|
|
}
|
|
|
|
/* raw bottom MB row */
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for(;s->mb_x < s->mb_width; s->mb_x++) {
|
|
ff_update_block_index(s);
|
|
vc1_put_signed_blocks_clamped(v);
|
|
if(v->s.loop_filter) vc1_loop_filter_iblk_delayed(v, v->pq);
|
|
}
|
|
if (v->s.loop_filter)
|
|
ff_draw_horiz_band(s, (s->mb_height-1)*16, 16);
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_width - 1, s->end_mb_y - 1, (AC_END|DC_END|MV_END));
|
|
}
|
|
|
|
static void vc1_decode_p_blocks(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
int apply_loop_filter;
|
|
|
|
/* select codingmode used for VLC tables selection */
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset = (v->pqindex <= 8) ? CS_HIGH_RATE_INTRA : CS_LOW_MOT_INTRA;
|
|
break;
|
|
case 1:
|
|
v->codingset = CS_HIGH_MOT_INTRA;
|
|
break;
|
|
case 2:
|
|
v->codingset = CS_MID_RATE_INTRA;
|
|
break;
|
|
}
|
|
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset2 = (v->pqindex <= 8) ? CS_HIGH_RATE_INTER : CS_LOW_MOT_INTER;
|
|
break;
|
|
case 1:
|
|
v->codingset2 = CS_HIGH_MOT_INTER;
|
|
break;
|
|
case 2:
|
|
v->codingset2 = CS_MID_RATE_INTER;
|
|
break;
|
|
}
|
|
|
|
apply_loop_filter = s->loop_filter && !(s->avctx->skip_loop_filter >= AVDISCARD_NONKEY);
|
|
s->first_slice_line = 1;
|
|
memset(v->cbp_base, 0, sizeof(v->cbp_base[0])*2*s->mb_stride);
|
|
for(s->mb_y = s->start_mb_y; s->mb_y < s->end_mb_y; s->mb_y++) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for(; s->mb_x < s->mb_width; s->mb_x++) {
|
|
ff_update_block_index(s);
|
|
|
|
vc1_decode_p_mb(v);
|
|
if (s->mb_y != s->start_mb_y && apply_loop_filter)
|
|
vc1_apply_p_loop_filter(v);
|
|
if(get_bits_count(&s->gb) > v->bits || get_bits_count(&s->gb) < 0) {
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_x, s->mb_y, (AC_END|DC_END|MV_END));
|
|
av_log(s->avctx, AV_LOG_ERROR, "Bits overconsumption: %i > %i at %ix%i\n", get_bits_count(&s->gb), v->bits,s->mb_x,s->mb_y);
|
|
return;
|
|
}
|
|
}
|
|
memmove(v->cbp_base, v->cbp, sizeof(v->cbp_base[0])*s->mb_stride);
|
|
memmove(v->ttblk_base, v->ttblk, sizeof(v->ttblk_base[0])*s->mb_stride);
|
|
memmove(v->is_intra_base, v->is_intra, sizeof(v->is_intra_base[0])*s->mb_stride);
|
|
memmove(v->luma_mv_base, v->luma_mv, sizeof(v->luma_mv_base[0])*s->mb_stride);
|
|
if (s->mb_y != s->start_mb_y) ff_draw_horiz_band(s, (s->mb_y-1) * 16, 16);
|
|
s->first_slice_line = 0;
|
|
}
|
|
if (apply_loop_filter) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for (; s->mb_x < s->mb_width; s->mb_x++) {
|
|
ff_update_block_index(s);
|
|
vc1_apply_p_loop_filter(v);
|
|
}
|
|
}
|
|
if (s->end_mb_y >= s->start_mb_y)
|
|
ff_draw_horiz_band(s, (s->end_mb_y-1) * 16, 16);
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_width - 1, s->end_mb_y - 1, (AC_END|DC_END|MV_END));
|
|
}
|
|
|
|
static void vc1_decode_b_blocks(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
|
|
/* select codingmode used for VLC tables selection */
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset = (v->pqindex <= 8) ? CS_HIGH_RATE_INTRA : CS_LOW_MOT_INTRA;
|
|
break;
|
|
case 1:
|
|
v->codingset = CS_HIGH_MOT_INTRA;
|
|
break;
|
|
case 2:
|
|
v->codingset = CS_MID_RATE_INTRA;
|
|
break;
|
|
}
|
|
|
|
switch(v->c_ac_table_index){
|
|
case 0:
|
|
v->codingset2 = (v->pqindex <= 8) ? CS_HIGH_RATE_INTER : CS_LOW_MOT_INTER;
|
|
break;
|
|
case 1:
|
|
v->codingset2 = CS_HIGH_MOT_INTER;
|
|
break;
|
|
case 2:
|
|
v->codingset2 = CS_MID_RATE_INTER;
|
|
break;
|
|
}
|
|
|
|
s->first_slice_line = 1;
|
|
for(s->mb_y = s->start_mb_y; s->mb_y < s->end_mb_y; s->mb_y++) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
for(; s->mb_x < s->mb_width; s->mb_x++) {
|
|
ff_update_block_index(s);
|
|
|
|
vc1_decode_b_mb(v);
|
|
if(get_bits_count(&s->gb) > v->bits || get_bits_count(&s->gb) < 0) {
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_x, s->mb_y, (AC_END|DC_END|MV_END));
|
|
av_log(s->avctx, AV_LOG_ERROR, "Bits overconsumption: %i > %i at %ix%i\n", get_bits_count(&s->gb), v->bits,s->mb_x,s->mb_y);
|
|
return;
|
|
}
|
|
if(v->s.loop_filter) vc1_loop_filter_iblk(v, v->pq);
|
|
}
|
|
if (!v->s.loop_filter)
|
|
ff_draw_horiz_band(s, s->mb_y * 16, 16);
|
|
else if (s->mb_y)
|
|
ff_draw_horiz_band(s, (s->mb_y-1) * 16, 16);
|
|
s->first_slice_line = 0;
|
|
}
|
|
if (v->s.loop_filter)
|
|
ff_draw_horiz_band(s, (s->mb_height-1)*16, 16);
|
|
ff_er_add_slice(s, 0, s->start_mb_y, s->mb_width - 1, s->end_mb_y - 1, (AC_END|DC_END|MV_END));
|
|
}
|
|
|
|
static void vc1_decode_skip_blocks(VC1Context *v)
|
|
{
|
|
MpegEncContext *s = &v->s;
|
|
|
|
ff_er_add_slice(s, 0, 0, s->mb_width - 1, s->mb_height - 1, (AC_END|DC_END|MV_END));
|
|
s->first_slice_line = 1;
|
|
for(s->mb_y = 0; s->mb_y < s->mb_height; s->mb_y++) {
|
|
s->mb_x = 0;
|
|
ff_init_block_index(s);
|
|
ff_update_block_index(s);
|
|
memcpy(s->dest[0], s->last_picture.data[0] + s->mb_y * 16 * s->linesize, s->linesize * 16);
|
|
memcpy(s->dest[1], s->last_picture.data[1] + s->mb_y * 8 * s->uvlinesize, s->uvlinesize * 8);
|
|
memcpy(s->dest[2], s->last_picture.data[2] + s->mb_y * 8 * s->uvlinesize, s->uvlinesize * 8);
|
|
ff_draw_horiz_band(s, s->mb_y * 16, 16);
|
|
s->first_slice_line = 0;
|
|
}
|
|
s->pict_type = AV_PICTURE_TYPE_P;
|
|
}
|
|
|
|
static void vc1_decode_blocks(VC1Context *v)
|
|
{
|
|
|
|
v->s.esc3_level_length = 0;
|
|
if(v->x8_type){
|
|
ff_intrax8_decode_picture(&v->x8, 2*v->pq+v->halfpq, v->pq*(!v->pquantizer) );
|
|
}else{
|
|
v->cur_blk_idx = 0;
|
|
v->left_blk_idx = -1;
|
|
v->topleft_blk_idx = 1;
|
|
v->top_blk_idx = 2;
|
|
switch(v->s.pict_type) {
|
|
case AV_PICTURE_TYPE_I:
|
|
if(v->profile == PROFILE_ADVANCED)
|
|
vc1_decode_i_blocks_adv(v);
|
|
else
|
|
vc1_decode_i_blocks(v);
|
|
break;
|
|
case AV_PICTURE_TYPE_P:
|
|
if(v->p_frame_skipped)
|
|
vc1_decode_skip_blocks(v);
|
|
else
|
|
vc1_decode_p_blocks(v);
|
|
break;
|
|
case AV_PICTURE_TYPE_B:
|
|
if(v->bi_type){
|
|
if(v->profile == PROFILE_ADVANCED)
|
|
vc1_decode_i_blocks_adv(v);
|
|
else
|
|
vc1_decode_i_blocks(v);
|
|
}else
|
|
vc1_decode_b_blocks(v);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static inline float get_float_val(GetBitContext* gb)
|
|
{
|
|
return (float)get_bits_long(gb, 30) / (1<<15) - (1<<14);
|
|
}
|
|
|
|
static void vc1_sprite_parse_transform(VC1Context *v, GetBitContext* gb, float c[7])
|
|
{
|
|
c[1] = c[3] = 0.0f;
|
|
|
|
switch (get_bits(gb, 2)) {
|
|
case 0:
|
|
c[0] = 1.0f;
|
|
c[2] = get_float_val(gb);
|
|
c[4] = 1.0f;
|
|
break;
|
|
case 1:
|
|
c[0] = c[4] = get_float_val(gb);
|
|
c[2] = get_float_val(gb);
|
|
break;
|
|
case 2:
|
|
c[0] = get_float_val(gb);
|
|
c[2] = get_float_val(gb);
|
|
c[4] = get_float_val(gb);
|
|
break;
|
|
case 3:
|
|
av_log_ask_for_sample(v->s.avctx, NULL);
|
|
c[0] = get_float_val(gb);
|
|
c[1] = get_float_val(gb);
|
|
c[2] = get_float_val(gb);
|
|
c[3] = get_float_val(gb);
|
|
c[4] = get_float_val(gb);
|
|
break;
|
|
}
|
|
c[5] = get_float_val(gb);
|
|
if (get_bits1(gb))
|
|
c[6] = get_float_val(gb);
|
|
else
|
|
c[6] = 1.0f;
|
|
}
|
|
|
|
static void vc1_parse_sprites(VC1Context *v, GetBitContext* gb)
|
|
{
|
|
int effect_type, effect_flag, effect_pcount1, effect_pcount2, i;
|
|
float effect_params1[14], effect_params2[10];
|
|
|
|
float coefs[2][7];
|
|
vc1_sprite_parse_transform(v, gb, coefs[0]);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "S1:");
|
|
for (i = 0; i < 7; i++)
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, " %.3f", coefs[0][i]);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "\n");
|
|
|
|
if (v->two_sprites) {
|
|
vc1_sprite_parse_transform(v, gb, coefs[1]);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "S2:");
|
|
for (i = 0; i < 7; i++)
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, " %.3f", coefs[1][i]);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "\n");
|
|
}
|
|
skip_bits(gb, 2);
|
|
if (effect_type = get_bits_long(gb, 30)){
|
|
switch (effect_pcount1 = get_bits(gb, 4)) {
|
|
case 2:
|
|
effect_params1[0] = get_float_val(gb);
|
|
effect_params1[1] = get_float_val(gb);
|
|
break;
|
|
case 7:
|
|
vc1_sprite_parse_transform(v, gb, effect_params1);
|
|
break;
|
|
case 14:
|
|
vc1_sprite_parse_transform(v, gb, effect_params1);
|
|
vc1_sprite_parse_transform(v, gb, &effect_params1[7]);
|
|
break;
|
|
default:
|
|
av_log_ask_for_sample(v->s.avctx, NULL);
|
|
return;
|
|
}
|
|
if (effect_type != 13 || effect_params1[0] != coefs[0][6]) {
|
|
// effect 13 is simple alpha blending and matches the opacity above
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "Effect: %d; params: ", effect_type);
|
|
for (i = 0; i < effect_pcount1; i++)
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, " %.3f", effect_params1[i]);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "\n");
|
|
}
|
|
|
|
effect_pcount2 = get_bits(gb, 16);
|
|
if (effect_pcount2 > 10) {
|
|
av_log(v->s.avctx, AV_LOG_ERROR, "Too many effect parameters\n");
|
|
return;
|
|
} else if (effect_pcount2) {
|
|
i = 0;
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "Effect params 2: ");
|
|
while (i < effect_pcount2){
|
|
effect_params2[i] = get_float_val(gb);
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, " %.3f", effect_params2[i]);
|
|
i++;
|
|
}
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "\n");
|
|
}
|
|
}
|
|
if (effect_flag = get_bits1(gb))
|
|
av_log(v->s.avctx, AV_LOG_DEBUG, "Effect flag set\n");
|
|
|
|
if (get_bits_count(gb) >= gb->size_in_bits +
|
|
(v->s.avctx->codec_id == CODEC_ID_WMV3 ? 64 : 0))
|
|
av_log(v->s.avctx, AV_LOG_ERROR, "Buffer overrun\n");
|
|
if (get_bits_count(gb) < gb->size_in_bits - 8)
|
|
av_log(v->s.avctx, AV_LOG_WARNING, "Buffer not fully read\n");
|
|
}
|
|
|
|
/** Initialize a VC1/WMV3 decoder
|
|
* @todo TODO: Handle VC-1 IDUs (Transport level?)
|
|
* @todo TODO: Decypher remaining bits in extra_data
|
|
*/
|
|
static av_cold int vc1_decode_init(AVCodecContext *avctx)
|
|
{
|
|
VC1Context *v = avctx->priv_data;
|
|
MpegEncContext *s = &v->s;
|
|
GetBitContext gb;
|
|
int i, cur_width, cur_height;
|
|
|
|
if (!avctx->extradata_size || !avctx->extradata) return -1;
|
|
if (!(avctx->flags & CODEC_FLAG_GRAY))
|
|
avctx->pix_fmt = avctx->get_format(avctx, avctx->codec->pix_fmts);
|
|
else
|
|
avctx->pix_fmt = PIX_FMT_GRAY8;
|
|
avctx->hwaccel = ff_find_hwaccel(avctx->codec->id, avctx->pix_fmt);
|
|
v->s.avctx = avctx;
|
|
avctx->flags |= CODEC_FLAG_EMU_EDGE;
|
|
v->s.flags |= CODEC_FLAG_EMU_EDGE;
|
|
|
|
if(avctx->idct_algo==FF_IDCT_AUTO){
|
|
avctx->idct_algo=FF_IDCT_WMV2;
|
|
}
|
|
|
|
if(ff_msmpeg4_decode_init(avctx) < 0)
|
|
return -1;
|
|
if (vc1_init_common(v) < 0) return -1;
|
|
ff_vc1dsp_init(&v->vc1dsp);
|
|
|
|
cur_width = avctx->coded_width = avctx->width;
|
|
cur_height = avctx->coded_height = avctx->height;
|
|
if (avctx->codec_id == CODEC_ID_WMV3)
|
|
{
|
|
int count = 0;
|
|
|
|
// looks like WMV3 has a sequence header stored in the extradata
|
|
// advanced sequence header may be before the first frame
|
|
// the last byte of the extradata is a version number, 1 for the
|
|
// samples we can decode
|
|
|
|
init_get_bits(&gb, avctx->extradata, avctx->extradata_size*8);
|
|
|
|
if (vc1_decode_sequence_header(avctx, v, &gb) < 0)
|
|
return -1;
|
|
|
|
count = avctx->extradata_size*8 - get_bits_count(&gb);
|
|
if (count>0)
|
|
{
|
|
av_log(avctx, AV_LOG_INFO, "Extra data: %i bits left, value: %X\n",
|
|
count, get_bits(&gb, count));
|
|
}
|
|
else if (count < 0)
|
|
{
|
|
av_log(avctx, AV_LOG_INFO, "Read %i bits in overflow\n", -count);
|
|
}
|
|
} else { // VC1/WVC1/WVP2
|
|
const uint8_t *start = avctx->extradata;
|
|
uint8_t *end = avctx->extradata + avctx->extradata_size;
|
|
const uint8_t *next;
|
|
int size, buf2_size;
|
|
uint8_t *buf2 = NULL;
|
|
int seq_initialized = 0, ep_initialized = 0;
|
|
|
|
if(avctx->extradata_size < 16) {
|
|
av_log(avctx, AV_LOG_ERROR, "Extradata size too small: %i\n", avctx->extradata_size);
|
|
return -1;
|
|
}
|
|
|
|
buf2 = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
|
start = find_next_marker(start, end); // in WVC1 extradata first byte is its size, but can be 0 in mkv
|
|
next = start;
|
|
for(; next < end; start = next){
|
|
next = find_next_marker(start + 4, end);
|
|
size = next - start - 4;
|
|
if(size <= 0) continue;
|
|
buf2_size = vc1_unescape_buffer(start + 4, size, buf2);
|
|
init_get_bits(&gb, buf2, buf2_size * 8);
|
|
switch(AV_RB32(start)){
|
|
case VC1_CODE_SEQHDR:
|
|
if(vc1_decode_sequence_header(avctx, v, &gb) < 0){
|
|
av_free(buf2);
|
|
return -1;
|
|
}
|
|
seq_initialized = 1;
|
|
break;
|
|
case VC1_CODE_ENTRYPOINT:
|
|
if(vc1_decode_entry_point(avctx, v, &gb) < 0){
|
|
av_free(buf2);
|
|
return -1;
|
|
}
|
|
ep_initialized = 1;
|
|
break;
|
|
}
|
|
}
|
|
av_free(buf2);
|
|
if(!seq_initialized || !ep_initialized){
|
|
av_log(avctx, AV_LOG_ERROR, "Incomplete extradata\n");
|
|
return -1;
|
|
}
|
|
v->res_sprite = (avctx->codec_tag == MKTAG('W','V','P','2'));
|
|
}
|
|
// Sequence header information may not have been parsed
|
|
// yet when ff_msmpeg4_decode_init was called the fist time
|
|
// above. If sequence information changes, we need to call
|
|
// it again.
|
|
if (cur_width != avctx->width ||
|
|
cur_height != avctx->height) {
|
|
MPV_common_end(s);
|
|
if(ff_msmpeg4_decode_init(avctx) < 0)
|
|
return -1;
|
|
avctx->coded_width = avctx->width;
|
|
avctx->coded_height = avctx->height;
|
|
}
|
|
|
|
avctx->profile = v->profile;
|
|
if (v->profile == PROFILE_ADVANCED)
|
|
avctx->level = v->level;
|
|
|
|
avctx->has_b_frames= !!(avctx->max_b_frames);
|
|
s->low_delay = !avctx->has_b_frames;
|
|
|
|
s->mb_width = (avctx->coded_width+15)>>4;
|
|
s->mb_height = (avctx->coded_height+15)>>4;
|
|
|
|
if (v->profile == PROFILE_ADVANCED || v->res_fasttx) {
|
|
for (i = 0; i < 64; i++) {
|
|
#define transpose(x) ((x>>3) | ((x&7)<<3))
|
|
v->zz_8x8[0][i] = transpose(wmv1_scantable[0][i]);
|
|
v->zz_8x8[1][i] = transpose(wmv1_scantable[1][i]);
|
|
v->zz_8x8[2][i] = transpose(wmv1_scantable[2][i]);
|
|
v->zz_8x8[3][i] = transpose(wmv1_scantable[3][i]);
|
|
}
|
|
v->left_blk_sh = 0;
|
|
v->top_blk_sh = 3;
|
|
} else {
|
|
memcpy(v->zz_8x8, wmv1_scantable, 4*64);
|
|
v->left_blk_sh = 3;
|
|
v->top_blk_sh = 0;
|
|
}
|
|
|
|
/* Allocate mb bitplanes */
|
|
v->mv_type_mb_plane = av_malloc(s->mb_stride * s->mb_height);
|
|
v->direct_mb_plane = av_malloc(s->mb_stride * s->mb_height);
|
|
v->acpred_plane = av_malloc(s->mb_stride * s->mb_height);
|
|
v->over_flags_plane = av_malloc(s->mb_stride * s->mb_height);
|
|
|
|
v->n_allocated_blks = s->mb_width + 2;
|
|
v->block = av_malloc(sizeof(*v->block) * v->n_allocated_blks);
|
|
v->cbp_base = av_malloc(sizeof(v->cbp_base[0]) * 2 * s->mb_stride);
|
|
v->cbp = v->cbp_base + s->mb_stride;
|
|
v->ttblk_base = av_malloc(sizeof(v->ttblk_base[0]) * 2 * s->mb_stride);
|
|
v->ttblk = v->ttblk_base + s->mb_stride;
|
|
v->is_intra_base = av_malloc(sizeof(v->is_intra_base[0]) * 2 * s->mb_stride);
|
|
v->is_intra = v->is_intra_base + s->mb_stride;
|
|
v->luma_mv_base = av_malloc(sizeof(v->luma_mv_base[0]) * 2 * s->mb_stride);
|
|
v->luma_mv = v->luma_mv_base + s->mb_stride;
|
|
|
|
/* allocate block type info in that way so it could be used with s->block_index[] */
|
|
v->mb_type_base = av_malloc(s->b8_stride * (s->mb_height * 2 + 1) + s->mb_stride * (s->mb_height + 1) * 2);
|
|
v->mb_type[0] = v->mb_type_base + s->b8_stride + 1;
|
|
v->mb_type[1] = v->mb_type_base + s->b8_stride * (s->mb_height * 2 + 1) + s->mb_stride + 1;
|
|
v->mb_type[2] = v->mb_type[1] + s->mb_stride * (s->mb_height + 1);
|
|
|
|
/* Init coded blocks info */
|
|
if (v->profile == PROFILE_ADVANCED)
|
|
{
|
|
// if (alloc_bitplane(&v->over_flags_plane, s->mb_width, s->mb_height) < 0)
|
|
// return -1;
|
|
// if (alloc_bitplane(&v->ac_pred_plane, s->mb_width, s->mb_height) < 0)
|
|
// return -1;
|
|
}
|
|
|
|
ff_intrax8_common_init(&v->x8,s);
|
|
return 0;
|
|
}
|
|
|
|
|
|
/** Decode a VC1/WMV3 frame
|
|
* @todo TODO: Handle VC-1 IDUs (Transport level?)
|
|
*/
|
|
static int vc1_decode_frame(AVCodecContext *avctx,
|
|
void *data, int *data_size,
|
|
AVPacket *avpkt)
|
|
{
|
|
const uint8_t *buf = avpkt->data;
|
|
int buf_size = avpkt->size, n_slices = 0, i;
|
|
VC1Context *v = avctx->priv_data;
|
|
MpegEncContext *s = &v->s;
|
|
AVFrame *pict = data;
|
|
uint8_t *buf2 = NULL;
|
|
const uint8_t *buf_start = buf;
|
|
struct {
|
|
uint8_t *buf;
|
|
GetBitContext gb;
|
|
int mby_start;
|
|
} *slices = NULL;
|
|
|
|
/* no supplementary picture */
|
|
if (buf_size == 0 || (buf_size == 4 && AV_RB32(buf) == VC1_CODE_ENDOFSEQ)) {
|
|
/* special case for last picture */
|
|
if (s->low_delay==0 && s->next_picture_ptr) {
|
|
*pict= *(AVFrame*)s->next_picture_ptr;
|
|
s->next_picture_ptr= NULL;
|
|
|
|
*data_size = sizeof(AVFrame);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* We need to set current_picture_ptr before reading the header,
|
|
* otherwise we cannot store anything in there. */
|
|
if(s->current_picture_ptr==NULL || s->current_picture_ptr->data[0]){
|
|
int i= ff_find_unused_picture(s, 0);
|
|
s->current_picture_ptr= &s->picture[i];
|
|
}
|
|
|
|
if (s->avctx->codec->capabilities&CODEC_CAP_HWACCEL_VDPAU){
|
|
if (v->profile < PROFILE_ADVANCED)
|
|
avctx->pix_fmt = PIX_FMT_VDPAU_WMV3;
|
|
else
|
|
avctx->pix_fmt = PIX_FMT_VDPAU_VC1;
|
|
}
|
|
|
|
//for advanced profile we may need to parse and unescape data
|
|
if (avctx->codec_id == CODEC_ID_VC1) {
|
|
int buf_size2 = 0;
|
|
buf2 = av_mallocz(buf_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
|
|
|
if(IS_MARKER(AV_RB32(buf))){ /* frame starts with marker and needs to be parsed */
|
|
const uint8_t *start, *end, *next;
|
|
int size;
|
|
|
|
next = buf;
|
|
for(start = buf, end = buf + buf_size; next < end; start = next){
|
|
next = find_next_marker(start + 4, end);
|
|
size = next - start - 4;
|
|
if(size <= 0) continue;
|
|
switch(AV_RB32(start)){
|
|
case VC1_CODE_FRAME:
|
|
if (avctx->hwaccel ||
|
|
s->avctx->codec->capabilities&CODEC_CAP_HWACCEL_VDPAU)
|
|
buf_start = start;
|
|
buf_size2 = vc1_unescape_buffer(start + 4, size, buf2);
|
|
break;
|
|
case VC1_CODE_ENTRYPOINT: /* it should be before frame data */
|
|
buf_size2 = vc1_unescape_buffer(start + 4, size, buf2);
|
|
init_get_bits(&s->gb, buf2, buf_size2*8);
|
|
vc1_decode_entry_point(avctx, v, &s->gb);
|
|
break;
|
|
case VC1_CODE_SLICE: {
|
|
int buf_size3;
|
|
slices = av_realloc(slices, sizeof(*slices) * (n_slices+1));
|
|
if (!slices) goto err;
|
|
slices[n_slices].buf = av_mallocz(buf_size + FF_INPUT_BUFFER_PADDING_SIZE);
|
|
if (!slices[n_slices].buf) goto err;
|
|
buf_size3 = vc1_unescape_buffer(start + 4, size,
|
|
slices[n_slices].buf);
|
|
init_get_bits(&slices[n_slices].gb, slices[n_slices].buf,
|
|
buf_size3 << 3);
|
|
slices[n_slices].mby_start = get_bits(&slices[n_slices].gb, 9);
|
|
n_slices++;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}else if(v->interlace && ((buf[0] & 0xC0) == 0xC0)){ /* WVC1 interlaced stores both fields divided by marker */
|
|
const uint8_t *divider;
|
|
|
|
divider = find_next_marker(buf, buf + buf_size);
|
|
if((divider == (buf + buf_size)) || AV_RB32(divider) != VC1_CODE_FIELD){
|
|
av_log(avctx, AV_LOG_ERROR, "Error in WVC1 interlaced frame\n");
|
|
goto err;
|
|
}
|
|
|
|
buf_size2 = vc1_unescape_buffer(buf, divider - buf, buf2);
|
|
// TODO
|
|
if(!v->warn_interlaced++)
|
|
av_log(v->s.avctx, AV_LOG_ERROR, "Interlaced WVC1 support is not implemented\n");
|
|
goto err;
|
|
}else{
|
|
buf_size2 = vc1_unescape_buffer(buf, buf_size, buf2);
|
|
}
|
|
init_get_bits(&s->gb, buf2, buf_size2*8);
|
|
} else
|
|
init_get_bits(&s->gb, buf, buf_size*8);
|
|
|
|
if (v->res_sprite) {
|
|
v->new_sprite = !get_bits1(&s->gb);
|
|
v->two_sprites = get_bits1(&s->gb);
|
|
if (!v->new_sprite)
|
|
goto end;
|
|
}
|
|
|
|
// do parse frame header
|
|
if(v->profile < PROFILE_ADVANCED) {
|
|
if(vc1_parse_frame_header(v, &s->gb) == -1) {
|
|
goto err;
|
|
}
|
|
} else {
|
|
if(vc1_parse_frame_header_adv(v, &s->gb) == -1) {
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
if (v->res_sprite && s->pict_type!=AV_PICTURE_TYPE_I) {
|
|
av_log(v->s.avctx, AV_LOG_WARNING, "Sprite decoder: expected I-frame\n");
|
|
}
|
|
|
|
s->current_picture_ptr->repeat_pict = 0;
|
|
if (v->rff){
|
|
s->current_picture_ptr->repeat_pict = 1;
|
|
}else if (v->rptfrm){
|
|
s->current_picture_ptr->repeat_pict = v->rptfrm * 2;
|
|
}
|
|
|
|
s->current_picture_ptr->top_field_first = v->tff;
|
|
|
|
// for skipping the frame
|
|
s->current_picture.pict_type= s->pict_type;
|
|
s->current_picture.key_frame= s->pict_type == AV_PICTURE_TYPE_I;
|
|
|
|
/* skip B-frames if we don't have reference frames */
|
|
if(s->last_picture_ptr==NULL && (s->pict_type==AV_PICTURE_TYPE_B || s->dropable)){
|
|
goto err;
|
|
}
|
|
#if FF_API_HURRY_UP
|
|
/* skip b frames if we are in a hurry */
|
|
if(avctx->hurry_up && s->pict_type==FF_B_TYPE) return -1;//buf_size;
|
|
#endif
|
|
if( (avctx->skip_frame >= AVDISCARD_NONREF && s->pict_type==AV_PICTURE_TYPE_B)
|
|
|| (avctx->skip_frame >= AVDISCARD_NONKEY && s->pict_type!=AV_PICTURE_TYPE_I)
|
|
|| avctx->skip_frame >= AVDISCARD_ALL) {
|
|
goto end;
|
|
}
|
|
#if FF_API_HURRY_UP
|
|
/* skip everything if we are in a hurry>=5 */
|
|
if(avctx->hurry_up>=5) {
|
|
goto err;
|
|
}
|
|
#endif
|
|
|
|
if(s->next_p_frame_damaged){
|
|
if(s->pict_type==AV_PICTURE_TYPE_B)
|
|
goto end;
|
|
else
|
|
s->next_p_frame_damaged=0;
|
|
}
|
|
|
|
if(MPV_frame_start(s, avctx) < 0) {
|
|
goto err;
|
|
}
|
|
|
|
s->me.qpel_put= s->dsp.put_qpel_pixels_tab;
|
|
s->me.qpel_avg= s->dsp.avg_qpel_pixels_tab;
|
|
|
|
if ((CONFIG_VC1_VDPAU_DECODER)
|
|
&&s->avctx->codec->capabilities&CODEC_CAP_HWACCEL_VDPAU)
|
|
ff_vdpau_vc1_decode_picture(s, buf_start, (buf + buf_size) - buf_start);
|
|
else if (avctx->hwaccel) {
|
|
if (avctx->hwaccel->start_frame(avctx, buf, buf_size) < 0)
|
|
goto err;
|
|
if (avctx->hwaccel->decode_slice(avctx, buf_start, (buf + buf_size) - buf_start) < 0)
|
|
goto err;
|
|
if (avctx->hwaccel->end_frame(avctx) < 0)
|
|
goto err;
|
|
} else {
|
|
ff_er_frame_start(s);
|
|
|
|
v->bits = buf_size * 8;
|
|
for (i = 0; i <= n_slices; i++) {
|
|
if (i && get_bits1(&s->gb))
|
|
vc1_parse_frame_header_adv(v, &s->gb);
|
|
s->start_mb_y = (i == 0) ? 0 : FFMAX(0, slices[i-1].mby_start);
|
|
s->end_mb_y = (i == n_slices) ? s->mb_height : FFMIN(s->mb_height, slices[i].mby_start);
|
|
vc1_decode_blocks(v);
|
|
if (i != n_slices) s->gb = slices[i].gb;
|
|
}
|
|
//av_log(s->avctx, AV_LOG_INFO, "Consumed %i/%i bits\n", get_bits_count(&s->gb), s->gb.size_in_bits);
|
|
// if(get_bits_count(&s->gb) > buf_size * 8)
|
|
// return -1;
|
|
ff_er_frame_end(s);
|
|
}
|
|
|
|
MPV_frame_end(s);
|
|
|
|
assert(s->current_picture.pict_type == s->current_picture_ptr->pict_type);
|
|
assert(s->current_picture.pict_type == s->pict_type);
|
|
if (s->pict_type == AV_PICTURE_TYPE_B || s->low_delay) {
|
|
*pict= *(AVFrame*)s->current_picture_ptr;
|
|
} else if (s->last_picture_ptr != NULL) {
|
|
*pict= *(AVFrame*)s->last_picture_ptr;
|
|
}
|
|
|
|
if(s->last_picture_ptr || s->low_delay){
|
|
*data_size = sizeof(AVFrame);
|
|
ff_print_debug_info(s, pict);
|
|
}
|
|
|
|
end:
|
|
if (v->res_sprite)
|
|
vc1_parse_sprites(v, &s->gb);
|
|
av_free(buf2);
|
|
for (i = 0; i < n_slices; i++)
|
|
av_free(slices[i].buf);
|
|
av_free(slices);
|
|
return buf_size;
|
|
|
|
err:
|
|
av_free(buf2);
|
|
for (i = 0; i < n_slices; i++)
|
|
av_free(slices[i].buf);
|
|
av_free(slices);
|
|
return -1;
|
|
}
|
|
|
|
|
|
/** Close a VC1/WMV3 decoder
|
|
* @warning Initial try at using MpegEncContext stuff
|
|
*/
|
|
static av_cold int vc1_decode_end(AVCodecContext *avctx)
|
|
{
|
|
VC1Context *v = avctx->priv_data;
|
|
|
|
av_freep(&v->hrd_rate);
|
|
av_freep(&v->hrd_buffer);
|
|
MPV_common_end(&v->s);
|
|
av_freep(&v->mv_type_mb_plane);
|
|
av_freep(&v->direct_mb_plane);
|
|
av_freep(&v->acpred_plane);
|
|
av_freep(&v->over_flags_plane);
|
|
av_freep(&v->mb_type_base);
|
|
av_freep(&v->block);
|
|
av_freep(&v->cbp_base);
|
|
av_freep(&v->ttblk_base);
|
|
av_freep(&v->is_intra_base); // FIXME use v->mb_type[]
|
|
av_freep(&v->luma_mv_base);
|
|
ff_intrax8_common_end(&v->x8);
|
|
return 0;
|
|
}
|
|
|
|
static const AVProfile profiles[] = {
|
|
{ FF_PROFILE_VC1_SIMPLE, "Simple" },
|
|
{ FF_PROFILE_VC1_MAIN, "Main" },
|
|
{ FF_PROFILE_VC1_COMPLEX, "Complex" },
|
|
{ FF_PROFILE_VC1_ADVANCED, "Advanced" },
|
|
{ FF_PROFILE_UNKNOWN },
|
|
};
|
|
|
|
AVCodec ff_vc1_decoder = {
|
|
"vc1",
|
|
AVMEDIA_TYPE_VIDEO,
|
|
CODEC_ID_VC1,
|
|
sizeof(VC1Context),
|
|
vc1_decode_init,
|
|
NULL,
|
|
vc1_decode_end,
|
|
vc1_decode_frame,
|
|
CODEC_CAP_DR1 | CODEC_CAP_DELAY,
|
|
NULL,
|
|
.long_name = NULL_IF_CONFIG_SMALL("SMPTE VC-1"),
|
|
.pix_fmts = ff_hwaccel_pixfmt_list_420,
|
|
.profiles = NULL_IF_CONFIG_SMALL(profiles)
|
|
};
|
|
|
|
#if CONFIG_WMV3_DECODER
|
|
AVCodec ff_wmv3_decoder = {
|
|
"wmv3",
|
|
AVMEDIA_TYPE_VIDEO,
|
|
CODEC_ID_WMV3,
|
|
sizeof(VC1Context),
|
|
vc1_decode_init,
|
|
NULL,
|
|
vc1_decode_end,
|
|
vc1_decode_frame,
|
|
CODEC_CAP_DR1 | CODEC_CAP_DELAY,
|
|
NULL,
|
|
.long_name = NULL_IF_CONFIG_SMALL("Windows Media Video 9"),
|
|
.pix_fmts = ff_hwaccel_pixfmt_list_420,
|
|
.profiles = NULL_IF_CONFIG_SMALL(profiles)
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_WMV3_VDPAU_DECODER
|
|
AVCodec ff_wmv3_vdpau_decoder = {
|
|
"wmv3_vdpau",
|
|
AVMEDIA_TYPE_VIDEO,
|
|
CODEC_ID_WMV3,
|
|
sizeof(VC1Context),
|
|
vc1_decode_init,
|
|
NULL,
|
|
vc1_decode_end,
|
|
vc1_decode_frame,
|
|
CODEC_CAP_DR1 | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU,
|
|
NULL,
|
|
.long_name = NULL_IF_CONFIG_SMALL("Windows Media Video 9 VDPAU"),
|
|
.pix_fmts = (const enum PixelFormat[]){PIX_FMT_VDPAU_WMV3, PIX_FMT_NONE},
|
|
.profiles = NULL_IF_CONFIG_SMALL(profiles)
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_VC1_VDPAU_DECODER
|
|
AVCodec ff_vc1_vdpau_decoder = {
|
|
"vc1_vdpau",
|
|
AVMEDIA_TYPE_VIDEO,
|
|
CODEC_ID_VC1,
|
|
sizeof(VC1Context),
|
|
vc1_decode_init,
|
|
NULL,
|
|
vc1_decode_end,
|
|
vc1_decode_frame,
|
|
CODEC_CAP_DR1 | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU,
|
|
NULL,
|
|
.long_name = NULL_IF_CONFIG_SMALL("SMPTE VC-1 VDPAU"),
|
|
.pix_fmts = (const enum PixelFormat[]){PIX_FMT_VDPAU_VC1, PIX_FMT_NONE},
|
|
.profiles = NULL_IF_CONFIG_SMALL(profiles)
|
|
};
|
|
#endif
|