[FFmpeg-cvslog] avcodec: loongson optimize h264dsp idct and loop filter with mmi

周晓勇 git at videolan.org
Mon Sep 7 23:03:02 CEST 2015


ffmpeg | branch: master | 周晓勇 <zhouxiaoyong at loongson.cn> | Wed Sep  2 18:02:24 2015 +0800| [f7e9b9d260f9c7f065791c70d9ca6b89b5164742] | committer: Michael Niedermayer

avcodec: loongson optimize h264dsp idct and loop filter with mmi

Change-Id: Ic87fb8f5cd22a502ff9dbbc5a5a8ea97cfc8a1dd
Signed-off-by: ZhouXiaoyong <zhouxiaoyong at loongson.cn>
Signed-off-by: Michael Niedermayer <michael at niedermayer.cc>

> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=f7e9b9d260f9c7f065791c70d9ca6b89b5164742
---

 libavcodec/mips/h264dsp_init_mips.c |   43 +-
 libavcodec/mips/h264dsp_mips.h      |   41 +
 libavcodec/mips/h264dsp_mmi.c       | 2222 +++++++++++++++++++++++++++++++++++
 3 files changed, 2303 insertions(+), 3 deletions(-)

diff --git a/libavcodec/mips/h264dsp_init_mips.c b/libavcodec/mips/h264dsp_init_mips.c
index 3694621..1fe7f84 100644
--- a/libavcodec/mips/h264dsp_init_mips.c
+++ b/libavcodec/mips/h264dsp_init_mips.c
@@ -82,11 +82,33 @@ static av_cold void h264dsp_init_msa(H264DSPContext *c,
 #endif  // #if HAVE_MSA
 
 #if HAVE_MMI
-static av_cold void h264dsp_init_mmi(H264DSPContext * c,
-                                     const int bit_depth,
-                                     const int chroma_format_idc)
+static av_cold void h264dsp_init_mmi(H264DSPContext * c, const int bit_depth,
+        const int chroma_format_idc)
 {
     if (bit_depth == 8) {
+        c->h264_add_pixels4_clear = ff_h264_add_pixels4_8_mmi;
+        c->h264_idct_add = ff_h264_idct_add_8_mmi;
+        c->h264_idct8_add = ff_h264_idct8_add_8_mmi;
+        c->h264_idct_dc_add = ff_h264_idct_dc_add_8_mmi;
+        c->h264_idct8_dc_add = ff_h264_idct8_dc_add_8_mmi;
+        c->h264_idct_add16 = ff_h264_idct_add16_8_mmi;
+        c->h264_idct_add16intra = ff_h264_idct_add16intra_8_mmi;
+        c->h264_idct8_add4 = ff_h264_idct8_add4_8_mmi;
+
+        if (chroma_format_idc <= 1)
+            c->h264_idct_add8 = ff_h264_idct_add8_8_mmi;
+        else
+            c->h264_idct_add8 = ff_h264_idct_add8_422_8_mmi;
+
+        c->h264_luma_dc_dequant_idct = ff_h264_luma_dc_dequant_idct_8_mmi;
+
+        if (chroma_format_idc <= 1)
+            c->h264_chroma_dc_dequant_idct =
+                ff_h264_chroma_dc_dequant_idct_8_mmi;
+        else
+            c->h264_chroma_dc_dequant_idct =
+                ff_h264_chroma422_dc_dequant_idct_8_mmi;
+
         c->weight_h264_pixels_tab[0] = ff_h264_weight_pixels16_8_mmi;
         c->weight_h264_pixels_tab[1] = ff_h264_weight_pixels8_8_mmi;
         c->weight_h264_pixels_tab[2] = ff_h264_weight_pixels4_8_mmi;
@@ -94,6 +116,21 @@ static av_cold void h264dsp_init_mmi(H264DSPContext * c,
         c->biweight_h264_pixels_tab[0] = ff_h264_biweight_pixels16_8_mmi;
         c->biweight_h264_pixels_tab[1] = ff_h264_biweight_pixels8_8_mmi;
         c->biweight_h264_pixels_tab[2] = ff_h264_biweight_pixels4_8_mmi;
+
+        c->h264_v_loop_filter_chroma       = ff_deblock_v_chroma_8_mmi;
+        c->h264_v_loop_filter_chroma_intra = ff_deblock_v_chroma_intra_8_mmi;
+
+        if (chroma_format_idc <= 1) {
+            c->h264_h_loop_filter_chroma =
+                ff_deblock_h_chroma_8_mmi;
+            c->h264_h_loop_filter_chroma_intra =
+                ff_deblock_h_chroma_intra_8_mmi;
+        }
+
+        c->h264_v_loop_filter_luma = ff_deblock_v_luma_8_mmi;
+        c->h264_v_loop_filter_luma_intra = ff_deblock_v_luma_intra_8_mmi;
+        c->h264_h_loop_filter_luma = ff_deblock_h_luma_8_mmi;
+        c->h264_h_loop_filter_luma_intra = ff_deblock_h_luma_intra_8_mmi;
     }
 }
 #endif /* HAVE_MMI */
diff --git a/libavcodec/mips/h264dsp_mips.h b/libavcodec/mips/h264dsp_mips.h
index 0b72c2e..3fdbf4f 100644
--- a/libavcodec/mips/h264dsp_mips.h
+++ b/libavcodec/mips/h264dsp_mips.h
@@ -319,6 +319,26 @@ void ff_vp8_pred8x8_129_dc_8_msa(uint8_t *src, ptrdiff_t stride);
 void ff_vp8_pred16x16_127_dc_8_msa(uint8_t *src, ptrdiff_t stride);
 void ff_vp8_pred16x16_129_dc_8_msa(uint8_t *src, ptrdiff_t stride);
 
+void ff_h264_add_pixels4_8_mmi(uint8_t *_dst, int16_t *_src, int stride);
+void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride);
+void ff_h264_idct8_add_8_mmi(uint8_t *dst, int16_t *block, int stride);
+void ff_h264_idct_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride);
+void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride);
+void ff_h264_idct_add16_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8]);
+void ff_h264_idct_add16intra_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8]);
+void ff_h264_idct8_add4_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8]);
+void ff_h264_idct_add8_8_mmi(uint8_t **dest, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8]);
+void ff_h264_idct_add8_422_8_mmi(uint8_t **dest, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8]);
+void ff_h264_luma_dc_dequant_idct_8_mmi(int16_t *output, int16_t *input,
+        int qmul);
+void ff_h264_chroma_dc_dequant_idct_8_mmi(int16_t *block, int qmul);
+void ff_h264_chroma422_dc_dequant_idct_8_mmi(int16_t *block, int qmul);
+
 void ff_h264_weight_pixels16_8_mmi(uint8_t *block, int stride, int height,
         int log2_denom, int weight, int offset);
 void ff_h264_biweight_pixels16_8_mmi(uint8_t *dst, uint8_t *src,
@@ -335,6 +355,27 @@ void ff_h264_biweight_pixels4_8_mmi(uint8_t *dst, uint8_t *src,
         int stride, int height, int log2_denom, int weightd, int weights,
         int offset);
 
+void ff_deblock_v_chroma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0);
+void ff_deblock_v_chroma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta);
+void ff_deblock_h_chroma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0);
+void ff_deblock_h_chroma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta);
+void ff_deblock_v_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0);
+void ff_deblock_v_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta);
+void ff_deblock_h_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0);
+void ff_deblock_h_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta);
+void ff_deblock_v8_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0);
+void ff_deblock_v8_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta);
+
 void ff_put_h264_qpel16_mc00_mmi(uint8_t *dst, const uint8_t *src,
         ptrdiff_t dst_stride);
 void ff_put_h264_qpel16_mc10_mmi(uint8_t *dst, const uint8_t *src,
diff --git a/libavcodec/mips/h264dsp_mmi.c b/libavcodec/mips/h264dsp_mmi.c
index 641cd2f..14c4a43 100644
--- a/libavcodec/mips/h264dsp_mmi.c
+++ b/libavcodec/mips/h264dsp_mmi.c
@@ -4,6 +4,7 @@
  * Copyright (c) 2015 Loongson Technology Corporation Limited
  * Copyright (c) 2015 Zhou Xiaoyong <zhouxiaoyong at loongson.cn>
  *                    Zhang Shuangshuang <zhangshuangshuang at ict.ac.cn>
+ *                    Heiher <r at hev.cc>
  *
  * This file is part of FFmpeg.
  *
@@ -25,6 +26,1011 @@
 #include "libavcodec/bit_depth_template.c"
 #include "h264dsp_mips.h"
 
+void ff_h264_add_pixels4_8_mmi(uint8_t *dst, int16_t *src, int stride)
+{
+    __asm__ volatile (
+        "xor $f0, $f0, $f0              \r\n"
+        "ldc1 $f2, 0(%[src])            \r\n"
+        "ldc1 $f4, 8(%[src])            \r\n"
+        "ldc1 $f6, 16(%[src])           \r\n"
+        "ldc1 $f8, 24(%[src])           \r\n"
+        "lwc1 $f10, 0(%[dst0])          \r\n"
+        "lwc1 $f12, 0(%[dst1])          \r\n"
+        "lwc1 $f14, 0(%[dst2])          \r\n"
+        "lwc1 $f16, 0(%[dst3])          \r\n"
+        "punpcklbh $f10, $f10, $f0      \r\n"
+        "punpcklbh $f12, $f12, $f0      \r\n"
+        "punpcklbh $f14, $f14, $f0      \r\n"
+        "punpcklbh $f16, $f16, $f0      \r\n"
+        "paddh $f2, $f2, $f10           \r\n"
+        "paddh $f4, $f4, $f12           \r\n"
+        "paddh $f6, $f6, $f14           \r\n"
+        "paddh $f8, $f8, $f16           \r\n"
+        "packushb $f2, $f2, $f0         \r\n"
+        "packushb $f4, $f4, $f0         \r\n"
+        "packushb $f6, $f6, $f0         \r\n"
+        "packushb $f8, $f8, $f0         \r\n"
+        "swc1 $f2, 0(%[dst0])           \r\n"
+        "swc1 $f4, 0(%[dst1])           \r\n"
+        "swc1 $f6, 0(%[dst2])           \r\n"
+        "swc1 $f8, 0(%[dst3])           \r\n"
+        ::[dst0]"r"(dst),[dst1]"r"(dst+stride),[dst2]"r"(dst+2*stride),
+          [dst3]"r"(dst+3*stride),[src]"r"(src)
+        : "$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14","$f16"
+    );
+
+    memset(src, 0, 32);
+}
+
+void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride)
+{
+    __asm__ volatile (
+        "dli $8, 1                              \r\n"
+        "ldc1 $f0, 0(%[block])                  \r\n"
+        "dmtc1 $8, $f16                         \r\n"
+        "ldc1 $f2, 8(%[block])                  \r\n"
+        "dli $8, 6                              \r\n"
+        "ldc1 $f4, 16(%[block])                 \r\n"
+        "dmtc1 $8, $f18                         \r\n"
+        "psrah $f8, $f2, $f16                   \r\n"
+        "ldc1 $f6, 24(%[block])                 \r\n"
+        "psrah $f10, $f6, $f16                  \r\n"
+        "psubh $f8, $f8, $f6                    \r\n"
+        "paddh $f10, $f10, $f2                  \r\n"
+        "paddh $f20, $f4, $f0                   \r\n"
+        "psubh $f0, $f0, $f4                    \r\n"
+        "paddh $f22, $f10, $f20                 \r\n"
+        "psubh $f4, $f20, $f10                  \r\n"
+        "paddh $f20, $f8, $f0                   \r\n"
+        "psubh $f0, $f0, $f8                    \r\n"
+        "punpckhhw $f2, $f22, $f20              \r\n"
+        "punpcklhw $f10, $f22, $f20             \r\n"
+        "punpckhhw $f8, $f0, $f4                \r\n"
+        "punpcklhw $f0, $f0, $f4                \r\n"
+        "punpckhwd $f4, $f10, $f0               \r\n"
+        "punpcklwd $f10, $f10, $f0              \r\n"
+        "punpcklwd $f20, $f2, $f8               \r\n"
+        "punpckhwd $f0, $f2, $f8                \r\n"
+        "paddh $f10, $f10, %[ff_pw_32]          \r\n"
+        "psrah $f8, $f4, $f16                   \r\n"
+        "psrah $f6, $f0, $f16                   \r\n"
+        "psubh $f8, $f8, $f0                    \r\n"
+        "paddh $f6, $f6, $f4                    \r\n"
+        "paddh $f2, $f20, $f10                  \r\n"
+        "psubh $f10, $f10, $f20                 \r\n"
+        "paddh $f20, $f6, $f2                   \r\n"
+        "psubh $f2, $f2, $f6                    \r\n"
+        "paddh $f22, $f8, $f10                  \r\n"
+        "xor $f14, $f14, $f14                   \r\n"
+        "psubh $f10, $f10, $f8                  \r\n"
+        "sdc1 $f14, 0(%[block])                 \r\n"
+        "sdc1 $f14, 8(%[block])                 \r\n"
+        "sdc1 $f14, 16(%[block])                \r\n"
+        "sdc1 $f14, 24(%[block])                \r\n"
+        "lwc1 $f4, 0(%[dst])                    \r\n"
+        "psrah $f6, $f20, $f18                  \r\n"
+        "gslwxc1 $f0, 0(%[dst], %[stride])      \r\n"
+        "psrah $f8, $f22, $f18                  \r\n"
+        "punpcklbh $f4, $f4, $f14               \r\n"
+        "punpcklbh $f0, $f0, $f14               \r\n"
+        "paddh $f4, $f4, $f6                    \r\n"
+        "paddh $f0, $f0, $f8                    \r\n"
+        "packushb $f4, $f4, $f14                \r\n"
+        "packushb $f0, $f0, $f14                \r\n"
+        "swc1 $f4, 0(%[dst])                    \r\n"
+        "gsswxc1 $f0, 0(%[dst], %[stride])      \r\n"
+        "daddu %[dst], %[dst], %[stride]        \r\n"
+        "daddu %[dst], %[dst], %[stride]        \r\n"
+        "lwc1 $f4, 0(%[dst])                    \r\n"
+        "psrah $f10, $f10, $f18                 \r\n"
+        "gslwxc1 $f0, 0(%[dst], %[stride])      \r\n"
+        "psrah $f2, $f2, $f18                   \r\n"
+        "punpcklbh $f4, $f4, $f14               \r\n"
+        "punpcklbh $f0, $f0, $f14               \r\n"
+        "paddh $f4, $f4, $f10                   \r\n"
+        "paddh $f0, $f0, $f2                    \r\n"
+        "packushb $f4, $f4, $f14                \r\n"
+        "swc1 $f4, 0(%[dst])                    \r\n"
+        "packushb $f0, $f0, $f14                \r\n"
+        "gsswxc1 $f0, 0(%[dst], %[stride])      \r\n"
+        ::[dst]"r"(dst),[block]"r"(block),[stride]"r"((uint64_t)stride),
+          [ff_pw_32]"f"(ff_pw_32)
+        : "$8","$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14","$f16",
+          "$f18","$f20","$f22"
+    );
+
+    memset(block, 0, 32);
+}
+
+void ff_h264_idct8_add_8_mmi(uint8_t *dst, int16_t *block, int stride)
+{
+    __asm__ volatile (
+        "lhu $10, 0x0(%[block])                     \r\n"
+        "daddiu $29, $29, -0x20                     \r\n"
+        "daddiu $10, $10, 0x20                      \r\n"
+        "ldc1 $f2, 0x10(%[block])                   \r\n"
+        "sh $10, 0x0(%[block])                      \r\n"
+        "ldc1 $f4, 0x20(%[block])                   \r\n"
+        "dli $10, 0x1                               \r\n"
+        "ldc1 $f6, 0x30(%[block])                   \r\n"
+        "dmtc1 $10, $f16                            \r\n"
+        "ldc1 $f10, 0x50(%[block])                  \r\n"
+        "ldc1 $f12, 0x60(%[block])                  \r\n"
+        "ldc1 $f14, 0x70(%[block])                  \r\n"
+        "mov.d $f0, $f2                             \r\n"
+        "psrah $f2, $f2, $f16                       \r\n"
+        "psrah $f8, $f10, $f16                      \r\n"
+        "paddh $f2, $f2, $f0                        \r\n"
+        "paddh $f8, $f8, $f10                       \r\n"
+        "paddh $f2, $f2, $f10                       \r\n"
+        "paddh $f8, $f8, $f14                       \r\n"
+        "paddh $f2, $f2, $f6                        \r\n"
+        "psubh $f8, $f8, $f0                        \r\n"
+        "psubh $f0, $f0, $f6                        \r\n"
+        "psubh $f10, $f10, $f6                      \r\n"
+        "psrah $f6, $f6, $f16                       \r\n"
+        "paddh $f0, $f0, $f14                       \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "psrah $f14, $f14, $f16                     \r\n"
+        "psubh $f0, $f0, $f6                        \r\n"
+        "dli $10, 0x2                               \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "dmtc1 $10, $f18                            \r\n"
+        "mov.d $f14, $f2                            \r\n"
+        "psrah $f2, $f2, $f18                       \r\n"
+        "psrah $f6, $f8, $f18                       \r\n"
+        "paddh $f6, $f6, $f0                        \r\n"
+        "psrah $f0, $f0, $f18                       \r\n"
+        "paddh $f2, $f2, $f10                       \r\n"
+        "psrah $f10, $f10, $f18                     \r\n"
+        "psubh $f0, $f0, $f8                        \r\n"
+        "psubh $f14, $f14, $f10                     \r\n"
+        "mov.d $f10, $f12                           \r\n"
+        "psrah $f12, $f12, $f16                     \r\n"
+        "psrah $f8, $f4, $f16                       \r\n"
+        "paddh $f12, $f12, $f4                      \r\n"
+        "psubh $f8, $f8, $f10                       \r\n"
+        "ldc1 $f4, 0x0(%[block])                    \r\n"
+        "ldc1 $f10, 0x40(%[block])                  \r\n"
+        "paddh $f10, $f10, $f4                      \r\n"
+        "paddh $f4, $f4, $f4                        \r\n"
+        "paddh $f12, $f12, $f10                     \r\n"
+        "psubh $f4, $f4, $f10                       \r\n"
+        "paddh $f10, $f10, $f10                     \r\n"
+        "paddh $f8, $f8, $f4                        \r\n"
+        "psubh $f10, $f10, $f12                     \r\n"
+        "paddh $f4, $f4, $f4                        \r\n"
+        "paddh $f14, $f14, $f12                     \r\n"
+        "psubh $f4, $f4, $f8                        \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f0, $f0, $f8                        \r\n"
+        "psubh $f12, $f12, $f14                     \r\n"
+        "paddh $f8, $f8, $f8                        \r\n"
+        "paddh $f6, $f6, $f4                        \r\n"
+        "psubh $f8, $f8, $f0                        \r\n"
+        "paddh $f4, $f4, $f4                        \r\n"
+        "paddh $f2, $f2, $f10                       \r\n"
+        "psubh $f4, $f4, $f6                        \r\n"
+        "paddh $f10, $f10, $f10                     \r\n"
+        "sdc1 $f12, 0x0(%[block])                   \r\n"
+        "psubh $f10, $f10, $f2                      \r\n"
+        "punpckhhw $f12, $f14, $f0                  \r\n"
+        "punpcklhw $f14, $f14, $f0                  \r\n"
+        "punpckhhw $f0, $f6, $f2                    \r\n"
+        "punpcklhw $f6, $f6, $f2                    \r\n"
+        "punpckhwd $f2, $f14, $f6                   \r\n"
+        "punpcklwd $f14, $f14, $f6                  \r\n"
+        "punpckhwd $f6, $f12, $f0                   \r\n"
+        "punpcklwd $f12, $f12, $f0                  \r\n"
+        "ldc1 $f0, 0x0(%[block])                    \r\n"
+        "sdc1 $f14, 0x0($29)                        \r\n"
+        "sdc1 $f2, 0x10($29)                        \r\n"
+        "dmfc1 $8, $f12                             \r\n"
+        "dmfc1 $11, $f6                             \r\n"
+        "punpckhhw $f6, $f10, $f4                   \r\n"
+        "punpcklhw $f10, $f10, $f4                  \r\n"
+        "punpckhhw $f4, $f8, $f0                    \r\n"
+        "punpcklhw $f8, $f8, $f0                    \r\n"
+        "punpckhwd $f0, $f10, $f8                   \r\n"
+        "punpcklwd $f10, $f10, $f8                  \r\n"
+        "punpckhwd $f8, $f6, $f4                    \r\n"
+        "punpcklwd $f6, $f6, $f4                    \r\n"
+        "sdc1 $f10, 0x8($29)                        \r\n"
+        "sdc1 $f0, 0x18($29)                        \r\n"
+        "dmfc1 $9, $f6                              \r\n"
+        "dmfc1 $12, $f8                             \r\n"
+        "ldc1 $f2, 0x18(%[block])                   \r\n"
+        "ldc1 $f12, 0x28(%[block])                  \r\n"
+        "ldc1 $f4, 0x38(%[block])                   \r\n"
+        "ldc1 $f0, 0x58(%[block])                   \r\n"
+        "ldc1 $f6, 0x68(%[block])                   \r\n"
+        "ldc1 $f8, 0x78(%[block])                   \r\n"
+        "mov.d $f14, $f2                            \r\n"
+        "psrah $f10, $f0, $f16                      \r\n"
+        "psrah $f2, $f2, $f16                       \r\n"
+        "paddh $f10, $f10, $f0                      \r\n"
+        "paddh $f2, $f2, $f14                       \r\n"
+        "paddh $f10, $f10, $f8                      \r\n"
+        "paddh $f2, $f2, $f0                        \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "paddh $f2, $f2, $f4                        \r\n"
+        "psubh $f14, $f14, $f4                      \r\n"
+        "psubh $f0, $f0, $f4                        \r\n"
+        "psrah $f4, $f4, $f16                       \r\n"
+        "paddh $f14, $f14, $f8                      \r\n"
+        "psubh $f0, $f0, $f8                        \r\n"
+        "psrah $f8, $f8, $f16                       \r\n"
+        "psubh $f14, $f14, $f4                      \r\n"
+        "psubh $f0, $f0, $f8                        \r\n"
+        "mov.d $f8, $f2                             \r\n"
+        "psrah $f4, $f10, $f18                      \r\n"
+        "psrah $f2, $f2, $f18                       \r\n"
+        "paddh $f4, $f4, $f14                       \r\n"
+        "psrah $f14, $f14, $f18                     \r\n"
+        "paddh $f2, $f2, $f0                        \r\n"
+        "psrah $f0, $f0, $f18                       \r\n"
+        "psubh $f14, $f14, $f10                     \r\n"
+        "psubh $f8, $f8, $f0                        \r\n"
+        "mov.d $f0, $f6                             \r\n"
+        "psrah $f6, $f6, $f16                       \r\n"
+        "psrah $f10, $f12, $f16                     \r\n"
+        "paddh $f6, $f6, $f12                       \r\n"
+        "psubh $f10, $f10, $f0                      \r\n"
+        "ldc1 $f12, 0x8(%[block])                   \r\n"
+        "ldc1 $f0, 0x48(%[block])                   \r\n"
+        "paddh $f0, $f0, $f12                       \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f6, $f6, $f0                        \r\n"
+        "psubh $f12, $f12, $f0                      \r\n"
+        "paddh $f0, $f0, $f0                        \r\n"
+        "paddh $f10, $f10, $f12                     \r\n"
+        "psubh $f0, $f0, $f6                        \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f8, $f8, $f6                        \r\n"
+        "psubh $f12, $f12, $f10                     \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "paddh $f14, $f14, $f10                     \r\n"
+        "psubh $f6, $f6, $f8                        \r\n"
+        "paddh $f10, $f10, $f10                     \r\n"
+        "paddh $f4, $f4, $f12                       \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f2, $f2, $f0                        \r\n"
+        "psubh $f12, $f12, $f4                      \r\n"
+        "paddh $f0, $f0, $f0                        \r\n"
+        "sdc1 $f6, 0x8(%[block])                    \r\n"
+        "psubh $f0, $f0, $f2                        \r\n"
+        "punpckhhw $f6, $f8, $f14                   \r\n"
+        "punpcklhw $f8, $f8, $f14                   \r\n"
+        "punpckhhw $f14, $f4, $f2                   \r\n"
+        "punpcklhw $f4, $f4, $f2                    \r\n"
+        "punpckhwd $f2, $f8, $f4                    \r\n"
+        "punpcklwd $f8, $f8, $f4                    \r\n"
+        "punpckhwd $f4, $f6, $f14                   \r\n"
+        "punpcklwd $f6, $f6, $f14                   \r\n"
+        "ldc1 $f14, 0x8(%[block])                   \r\n"
+        "dmfc1 $13, $f8                             \r\n"
+        "dmfc1 $15, $f2                             \r\n"
+        "mov.d $f24, $f6                            \r\n"
+        "mov.d $f28, $f4                            \r\n"
+        "punpckhhw $f4, $f0, $f12                   \r\n"
+        "punpcklhw $f0, $f0, $f12                   \r\n"
+        "punpckhhw $f12, $f10, $f14                 \r\n"
+        "punpcklhw $f10, $f10, $f14                 \r\n"
+        "punpckhwd $f14, $f0, $f10                  \r\n"
+        "punpcklwd $f0, $f0, $f10                   \r\n"
+        "punpckhwd $f10, $f4, $f12                  \r\n"
+        "punpcklwd $f4, $f4, $f12                   \r\n"
+        "dmfc1 $14, $f0                             \r\n"
+        "mov.d $f22, $f14                           \r\n"
+        "mov.d $f26, $f4                            \r\n"
+        "mov.d $f30, $f10                           \r\n"
+        "daddiu $10, %[dst], 0x4                    \r\n"
+        "dmtc1 $15, $f14                            \r\n"
+        "dmtc1 $11, $f12                            \r\n"
+        "ldc1 $f2, 0x10($29)                        \r\n"
+        "dmtc1 $8, $f6                              \r\n"
+        "mov.d $f8, $f2                             \r\n"
+        "psrah $f2, $f2, $f16                       \r\n"
+        "psrah $f0, $f14, $f16                      \r\n"
+        "paddh $f2, $f2, $f8                        \r\n"
+        "paddh $f0, $f0, $f14                       \r\n"
+        "paddh $f2, $f2, $f14                       \r\n"
+        "paddh $f0, $f0, $f28                       \r\n"
+        "paddh $f2, $f2, $f12                       \r\n"
+        "psubh $f0, $f0, $f8                        \r\n"
+        "psubh $f8, $f8, $f12                       \r\n"
+        "psubh $f14, $f14, $f12                     \r\n"
+        "psrah $f12, $f12, $f16                     \r\n"
+        "paddh $f8, $f8, $f28                       \r\n"
+        "psubh $f14, $f14, $f28                     \r\n"
+        "psrah $f10, $f28, $f16                     \r\n"
+        "psubh $f8, $f8, $f12                       \r\n"
+        "psubh $f14, $f14, $f10                     \r\n"
+        "mov.d $f10, $f2                            \r\n"
+        "psrah $f2, $f2, $f18                       \r\n"
+        "psrah $f12, $f0, $f18                      \r\n"
+        "paddh $f2, $f2, $f14                       \r\n"
+        "paddh $f12, $f12, $f8                      \r\n"
+        "psrah $f8, $f8, $f18                       \r\n"
+        "psrah $f14, $f14, $f18                     \r\n"
+        "psubh $f8, $f8, $f0                        \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "mov.d $f14, $f24                           \r\n"
+        "psrah $f4, $f24, $f16                      \r\n"
+        "psrah $f0, $f6, $f16                       \r\n"
+        "paddh $f4, $f4, $f6                        \r\n"
+        "psubh $f0, $f0, $f14                       \r\n"
+        "ldc1 $f6, 0x0($29)                         \r\n"
+        "dmtc1 $13, $f14                            \r\n"
+        "paddh $f14, $f14, $f6                      \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "paddh $f4, $f4, $f14                       \r\n"
+        "psubh $f6, $f6, $f14                       \r\n"
+        "paddh $f14, $f14, $f14                     \r\n"
+        "paddh $f0, $f0, $f6                        \r\n"
+        "psubh $f14, $f14, $f4                      \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "paddh $f10, $f10, $f4                      \r\n"
+        "psubh $f6, $f6, $f0                        \r\n"
+        "paddh $f4, $f4, $f4                        \r\n"
+        "paddh $f8, $f8, $f0                        \r\n"
+        "psubh $f4, $f4, $f10                       \r\n"
+        "paddh $f0, $f0, $f0                        \r\n"
+        "paddh $f12, $f12, $f6                      \r\n"
+        "psubh $f0, $f0, $f8                        \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "paddh $f2, $f2, $f14                       \r\n"
+        "psubh $f6, $f6, $f12                       \r\n"
+        "paddh $f14, $f14, $f14                     \r\n"
+        "sdc1 $f6, 0x0($29)                         \r\n"
+        "psubh $f14, $f14, $f2                      \r\n"
+        "sdc1 $f0, 0x10($29)                        \r\n"
+        "dmfc1 $8, $f4                              \r\n"
+        "xor $f4, $f4, $f4                          \r\n"
+        "sdc1 $f4, 0x0(%[block])                    \r\n"
+        "sdc1 $f4, 0x8(%[block])                    \r\n"
+        "sdc1 $f4, 0x10(%[block])                   \r\n"
+        "sdc1 $f4, 0x18(%[block])                   \r\n"
+        "sdc1 $f4, 0x20(%[block])                   \r\n"
+        "sdc1 $f4, 0x28(%[block])                   \r\n"
+        "sdc1 $f4, 0x30(%[block])                   \r\n"
+        "sdc1 $f4, 0x38(%[block])                   \r\n"
+        "sdc1 $f4, 0x40(%[block])                   \r\n"
+        "sdc1 $f4, 0x48(%[block])                   \r\n"
+        "sdc1 $f4, 0x50(%[block])                   \r\n"
+        "sdc1 $f4, 0x58(%[block])                   \r\n"
+        "sdc1 $f4, 0x60(%[block])                   \r\n"
+        "sdc1 $f4, 0x68(%[block])                   \r\n"
+        "sdc1 $f4, 0x70(%[block])                   \r\n"
+        "sdc1 $f4, 0x78(%[block])                   \r\n"
+        "dli $11, 0x6                               \r\n"
+        "lwc1 $f6, 0x0(%[dst])                      \r\n"
+        "dmtc1 $11, $f20                            \r\n"
+        "gslwxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "psrah $f10, $f10, $f20                     \r\n"
+        "psrah $f8, $f8, $f20                       \r\n"
+        "punpcklbh $f6, $f6, $f4                    \r\n"
+        "punpcklbh $f0, $f0, $f4                    \r\n"
+        "paddh $f6, $f6, $f10                       \r\n"
+        "paddh $f0, $f0, $f8                        \r\n"
+        "packushb $f6, $f6, $f4                     \r\n"
+        "packushb $f0, $f0, $f4                     \r\n"
+        "swc1 $f6, 0x0(%[dst])                      \r\n"
+        "gsswxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "lwc1 $f6, 0x0(%[dst])                      \r\n"
+        "gslwxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "psrah $f12, $f12, $f20                     \r\n"
+        "psrah $f2, $f2, $f20                       \r\n"
+        "punpcklbh $f6, $f6, $f4                    \r\n"
+        "punpcklbh $f0, $f0, $f4                    \r\n"
+        "paddh $f6, $f6, $f12                       \r\n"
+        "paddh $f0, $f0, $f2                        \r\n"
+        "packushb $f6, $f6, $f4                     \r\n"
+        "packushb $f0, $f0, $f4                     \r\n"
+        "swc1 $f6, 0x0(%[dst])                      \r\n"
+        "gsswxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "ldc1 $f10, 0x0($29)                        \r\n"
+        "ldc1 $f8, 0x10($29)                        \r\n"
+        "dmtc1 $8, $f12                             \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "lwc1 $f6, 0x0(%[dst])                      \r\n"
+        "gslwxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "psrah $f14, $f14, $f20                     \r\n"
+        "psrah $f10, $f10, $f20                     \r\n"
+        "punpcklbh $f6, $f6, $f4                    \r\n"
+        "punpcklbh $f0, $f0, $f4                    \r\n"
+        "paddh $f6, $f6, $f14                       \r\n"
+        "paddh $f0, $f0, $f10                       \r\n"
+        "packushb $f6, $f6, $f4                     \r\n"
+        "packushb $f0, $f0, $f4                     \r\n"
+        "swc1 $f6, 0x0(%[dst])                      \r\n"
+        "gsswxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "daddu %[dst], %[dst], %[stride]            \r\n"
+        "lwc1 $f6, 0x0(%[dst])                      \r\n"
+        "gslwxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "psrah $f8, $f8, $f20                       \r\n"
+        "psrah $f12, $f12, $f20                     \r\n"
+        "punpcklbh $f6, $f6, $f4                    \r\n"
+        "punpcklbh $f0, $f0, $f4                    \r\n"
+        "paddh $f6, $f6, $f8                        \r\n"
+        "paddh $f0, $f0, $f12                       \r\n"
+        "packushb $f6, $f6, $f4                     \r\n"
+        "packushb $f0, $f0, $f4                     \r\n"
+        "swc1 $f6, 0x0(%[dst])                      \r\n"
+        "gsswxc1 $f0, 0x0(%[dst], %[stride])        \r\n"
+        "dmtc1 $12, $f2                             \r\n"
+        "dmtc1 $9, $f12                             \r\n"
+        "ldc1 $f8, 0x18($29)                        \r\n"
+        "mov.d $f10, $f8                            \r\n"
+        "psrah $f8, $f8, $f16                       \r\n"
+        "psrah $f14, $f22, $f16                     \r\n"
+        "paddh $f14, $f14, $f22                     \r\n"
+        "paddh $f8, $f8, $f10                       \r\n"
+        "paddh $f14, $f14, $f30                     \r\n"
+        "paddh $f8, $f8, $f22                       \r\n"
+        "psubh $f14, $f14, $f10                     \r\n"
+        "paddh $f8, $f8, $f2                        \r\n"
+        "psubh $f10, $f10, $f2                      \r\n"
+        "psubh $f6, $f22, $f2                       \r\n"
+        "psrah $f2, $f2, $f16                       \r\n"
+        "paddh $f10, $f10, $f30                     \r\n"
+        "psubh $f6, $f6, $f30                       \r\n"
+        "psrah $f4, $f30, $f16                      \r\n"
+        "psubh $f10, $f10, $f2                      \r\n"
+        "psubh $f6, $f6, $f4                        \r\n"
+        "mov.d $f4, $f8                             \r\n"
+        "psrah $f8, $f8, $f18                       \r\n"
+        "psrah $f2, $f14, $f18                      \r\n"
+        "paddh $f8, $f8, $f6                        \r\n"
+        "paddh $f2, $f2, $f10                       \r\n"
+        "psrah $f10, $f10, $f18                     \r\n"
+        "psrah $f6, $f6, $f18                       \r\n"
+        "psubh $f10, $f10, $f14                     \r\n"
+        "psubh $f4, $f4, $f6                        \r\n"
+        "mov.d $f6, $f26                            \r\n"
+        "psrah $f0, $f26, $f16                      \r\n"
+        "psrah $f14, $f12, $f16                     \r\n"
+        "paddh $f0, $f0, $f12                       \r\n"
+        "psubh $f14, $f14, $f6                      \r\n"
+        "ldc1 $f12, 0x8($29)                        \r\n"
+        "dmtc1 $14, $f6                             \r\n"
+        "paddh $f6, $f6, $f12                       \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f0, $f0, $f6                        \r\n"
+        "psubh $f12, $f12, $f6                      \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "paddh $f14, $f14, $f12                     \r\n"
+        "psubh $f6, $f6, $f0                        \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f4, $f4, $f0                        \r\n"
+        "psubh $f12, $f12, $f14                     \r\n"
+        "paddh $f0, $f0, $f0                        \r\n"
+        "paddh $f10, $f10, $f14                     \r\n"
+        "psubh $f0, $f0, $f4                        \r\n"
+        "paddh $f14, $f14, $f14                     \r\n"
+        "paddh $f2, $f2, $f12                       \r\n"
+        "psubh $f14, $f14, $f10                     \r\n"
+        "paddh $f12, $f12, $f12                     \r\n"
+        "paddh $f8, $f8, $f6                        \r\n"
+        "psubh $f12, $f12, $f2                      \r\n"
+        "paddh $f6, $f6, $f6                        \r\n"
+        "sdc1 $f12, 0x8($29)                        \r\n"
+        "psubh $f6, $f6, $f8                        \r\n"
+        "sdc1 $f14, 0x18($29)                       \r\n"
+        "dmfc1 $9, $f0                              \r\n"
+        "xor $f0, $f0, $f0                          \r\n"
+        "lwc1 $f12, 0x0($10)                        \r\n"
+        "gslwxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "psrah $f4, $f4, $f20                       \r\n"
+        "psrah $f10, $f10, $f20                     \r\n"
+        "punpcklbh $f12, $f12, $f0                  \r\n"
+        "punpcklbh $f14, $f14, $f0                  \r\n"
+        "paddh $f12, $f12, $f4                      \r\n"
+        "paddh $f14, $f14, $f10                     \r\n"
+        "packushb $f12, $f12, $f0                   \r\n"
+        "packushb $f14, $f14, $f0                   \r\n"
+        "swc1 $f12, 0x0($10)                        \r\n"
+        "gsswxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "lwc1 $f12, 0x0($10)                        \r\n"
+        "gslwxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "psrah $f2, $f2, $f20                       \r\n"
+        "psrah $f8, $f8, $f20                       \r\n"
+        "punpcklbh $f12, $f12, $f0                  \r\n"
+        "punpcklbh $f14, $f14, $f0                  \r\n"
+        "paddh $f12, $f12, $f2                      \r\n"
+        "paddh $f14, $f14, $f8                      \r\n"
+        "packushb $f12, $f12, $f0                   \r\n"
+        "packushb $f14, $f14, $f0                   \r\n"
+        "swc1 $f12, 0x0($10)                        \r\n"
+        "gsswxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "ldc1 $f4, 0x8($29)                         \r\n"
+        "ldc1 $f10, 0x18($29)                       \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "dmtc1 $9, $f2                              \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "lwc1 $f12, 0x0($10)                        \r\n"
+        "gslwxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "psrah $f6, $f6, $f20                       \r\n"
+        "psrah $f4, $f4, $f20                       \r\n"
+        "punpcklbh $f12, $f12, $f0                  \r\n"
+        "punpcklbh $f14, $f14, $f0                  \r\n"
+        "paddh $f12, $f12, $f6                      \r\n"
+        "paddh $f14, $f14, $f4                      \r\n"
+        "packushb $f12, $f12, $f0                   \r\n"
+        "packushb $f14, $f14, $f0                   \r\n"
+        "swc1 $f12, 0x0($10)                        \r\n"
+        "gsswxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "daddu $10, $10, %[stride]                  \r\n"
+        "lwc1 $f12, 0x0($10)                        \r\n"
+        "gslwxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "psrah $f10, $f10, $f20                     \r\n"
+        "psrah $f2, $f2, $f20                       \r\n"
+        "punpcklbh $f12, $f12, $f0                  \r\n"
+        "punpcklbh $f14, $f14, $f0                  \r\n"
+        "paddh $f12, $f12, $f10                     \r\n"
+        "paddh $f14, $f14, $f2                      \r\n"
+        "packushb $f12, $f12, $f0                   \r\n"
+        "packushb $f14, $f14, $f0                   \r\n"
+        "swc1 $f12, 0x0($10)                        \r\n"
+        "gsswxc1 $f14, 0x0($10, %[stride])          \r\n"
+        "daddiu $29, $29, 0x20                      \r\n"
+        ::[dst]"r"(dst),[block]"r"(block),[stride]"r"((uint64_t)stride)
+        :"$8","$9","$10","$11","$12","$13","$14","$15","$29","$f0","$f2","$f4",
+         "$f8","$f10","$f12","$f14","$f16","$f18","$f20","$f22","$f24","$f26",
+         "$f28","$f30"
+    );
+
+    memset(block, 0, 128);
+}
+
+void ff_h264_idct_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride)
+{
+    __asm__ volatile (
+        "lh $8, 0x0(%[block])                       \r\n"
+        "sd $0, 0x0(%[block])                       \r\n"
+        "daddiu $8, $8, 0x20                        \r\n"
+        "daddu $10, %[stride], %[stride]            \r\n"
+        "dsra $8, $8, 0x6                           \r\n"
+        "xor $f2, $f2, $f2                          \r\n"
+        "mtc1 $8, $f0                               \r\n"
+        "pshufh $f0, $f0, $f2                       \r\n"
+        "daddu $8, $10, %[stride]                   \r\n"
+        "psubh $f2, $f2, $f0                        \r\n"
+        "packushb $f0, $f0, $f0                     \r\n"
+        "packushb $f2, $f2, $f2                     \r\n"
+        "lwc1 $f4, 0x0(%[dst])                      \r\n"
+        "gslwxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gslwxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "gslwxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        "paddusb $f4, $f4, $f0                      \r\n"
+        "paddusb $f6, $f6, $f0                      \r\n"
+        "paddusb $f8, $f8, $f0                      \r\n"
+        "paddusb $f10, $f10, $f0                    \r\n"
+        "psubusb $f4, $f4, $f2                      \r\n"
+        "psubusb $f6, $f6, $f2                      \r\n"
+        "psubusb $f8, $f8, $f2                      \r\n"
+        "psubusb $f10, $f10, $f2                    \r\n"
+        "swc1 $f4, 0x0(%[dst])                      \r\n"
+        "gsswxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gsswxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "gsswxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        ::[dst]"r"(dst),[block]"r"(block),[stride]"r"((uint64_t)stride)
+        : "$8","$10","$f0","$f2","$f4","$f6","$f8","$f10"
+    );
+}
+
+void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride)
+{
+    __asm__ volatile (
+        "lh $8, 0x0(%[block])                       \r\n"
+        "sd $0, 0x0(%[block])                       \r\n"
+        "daddiu $8, $8, 0x20                        \r\n"
+        "daddu $10, %[stride], %[stride]            \r\n"
+        "dsra $8, $8, 0x6                           \r\n"
+        "xor $f2, $f2, $f2                          \r\n"
+        "mtc1 $8, $f0                               \r\n"
+        "pshufh $f0, $f0, $f2                       \r\n"
+        "daddu $8, $10, %[stride]                   \r\n"
+        "psubh $f2, $f2, $f0                        \r\n"
+        "packushb $f0, $f0, $f0                     \r\n"
+        "packushb $f2, $f2, $f2                     \r\n"
+        "ldc1 $f4, 0x0(%[dst])                      \r\n"
+        "gsldxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gsldxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "gsldxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        "paddusb $f4, $f4, $f0                      \r\n"
+        "paddusb $f6, $f6, $f0                      \r\n"
+        "paddusb $f8, $f8, $f0                      \r\n"
+        "paddusb $f10, $f10, $f0                    \r\n"
+        "psubusb $f4, $f4, $f2                      \r\n"
+        "psubusb $f6, $f6, $f2                      \r\n"
+        "psubusb $f8, $f8, $f2                      \r\n"
+        "psubusb $f10, $f10, $f2                    \r\n"
+        "sdc1 $f4, 0x0(%[dst])                      \r\n"
+        "gssdxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gssdxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "daddu $9, $10, $10                         \r\n"
+        "gssdxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        "daddu %[dst], %[dst], $9                   \r\n"
+        "ldc1 $f4, 0x0(%[dst])                      \r\n"
+        "gsldxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gsldxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "gsldxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        "paddusb $f4, $f4, $f0                      \r\n"
+        "paddusb $f6, $f6, $f0                      \r\n"
+        "paddusb $f8, $f8, $f0                      \r\n"
+        "paddusb $f10, $f10, $f0                    \r\n"
+        "psubusb $f4, $f4, $f2                      \r\n"
+        "psubusb $f6, $f6, $f2                      \r\n"
+        "psubusb $f8, $f8, $f2                      \r\n"
+        "psubusb $f10, $f10, $f2                    \r\n"
+        "sdc1 $f4, 0x0(%[dst])                      \r\n"
+        "gssdxc1 $f6, 0x0(%[dst], %[stride])        \r\n"
+        "gssdxc1 $f8, 0x0(%[dst], $10)              \r\n"
+        "gssdxc1 $f10, 0x0(%[dst], $8)              \r\n"
+        ::[dst]"r"(dst),[block]"r"(block),[stride]"r"((uint64_t)stride)
+        : "$8","$9","$10","$f0","$f2","$f4","$f6","$f8","$f10"
+    );
+}
+
+void ff_h264_idct_add16_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8])
+{
+    int i;
+    for(i=0; i<16; i++){
+        int nnz = nnzc[ scan8[i] ];
+        if(nnz){
+            if(nnz==1 && ((int16_t*)block)[i*16])
+                ff_h264_idct_dc_add_8_mmi(dst + block_offset[i], block + i*16,
+                        stride);
+            else
+                ff_h264_idct_add_8_mmi(dst + block_offset[i], block + i*16,
+                        stride);
+        }
+    }
+}
+
+void ff_h264_idct_add16intra_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8])
+{
+    int i;
+    for(i=0; i<16; i++){
+        if(nnzc[ scan8[i] ])
+            ff_h264_idct_add_8_mmi(dst + block_offset[i], block + i*16, stride);
+        else if(((int16_t*)block)[i*16])
+            ff_h264_idct_dc_add_8_mmi(dst + block_offset[i], block + i*16,
+                    stride);
+    }
+}
+
+void ff_h264_idct8_add4_8_mmi(uint8_t *dst, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8])
+{
+    int i;
+    for(i=0; i<16; i+=4){
+        int nnz = nnzc[ scan8[i] ];
+        if(nnz){
+            if(nnz==1 && ((int16_t*)block)[i*16])
+                ff_h264_idct8_dc_add_8_mmi(dst + block_offset[i],
+                        block + i*16, stride);
+            else
+                ff_h264_idct8_add_8_mmi(dst + block_offset[i], block + i*16,
+                        stride);
+        }
+    }
+}
+
+void ff_h264_idct_add8_8_mmi(uint8_t **dest, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8])
+{
+    int i, j;
+    for(j=1; j<3; j++){
+        for(i=j*16; i<j*16+4; i++){
+            if(nnzc[ scan8[i] ])
+                ff_h264_idct_add_8_mmi(dest[j-1] + block_offset[i],
+                        block + i*16, stride);
+            else if(((int16_t*)block)[i*16])
+                ff_h264_idct_dc_add_8_mmi(dest[j-1] + block_offset[i],
+                        block + i*16, stride);
+        }
+    }
+}
+
+void ff_h264_idct_add8_422_8_mmi(uint8_t **dest, const int *block_offset,
+        int16_t *block, int stride, const uint8_t nnzc[15*8])
+{
+    int i, j;
+
+    for(j=1; j<3; j++){
+        for(i=j*16; i<j*16+4; i++){
+            if(nnzc[ scan8[i] ])
+                ff_h264_idct_add_8_mmi(dest[j-1] + block_offset[i],
+                        block + i*16, stride);
+            else if(((int16_t*)block)[i*16])
+                ff_h264_idct_dc_add_8_mmi(dest[j-1] + block_offset[i],
+                        block + i*16, stride);
+        }
+    }
+
+    for(j=1; j<3; j++){
+        for(i=j*16+4; i<j*16+8; i++){
+            if(nnzc[ scan8[i+4] ])
+                ff_h264_idct_add_8_mmi(dest[j-1] + block_offset[i+4],
+                        block + i*16, stride);
+            else if(((int16_t*)block)[i*16])
+                ff_h264_idct_dc_add_8_mmi(dest[j-1] + block_offset[i+4],
+                        block + i*16, stride);
+        }
+    }
+}
+
+void ff_h264_luma_dc_dequant_idct_8_mmi(int16_t *output, int16_t *input,
+        int qmul)
+{
+    __asm__ volatile (
+        ".set noreorder                                 \r\n"
+        "dli $10, 0x8                                   \r\n"
+        "ldc1 $f6, 0x18(%[input])                       \r\n"
+        "dmtc1 $10, $f16                                \r\n"
+        "ldc1 $f4, 0x10(%[input])                       \r\n"
+        "dli $10, 0x20                                  \r\n"
+        "ldc1 $f2, 0x8(%[input])                        \r\n"
+        "dmtc1 $10, $f18                                \r\n"
+        "ldc1 $f0, 0x0(%[input])                        \r\n"
+        "mov.d $f8, $f6                                 \r\n"
+        "paddh $f6, $f6, $f4                            \r\n"
+        "psubh $f4, $f4, $f8                            \r\n"
+        "mov.d $f8, $f2                                 \r\n"
+        "paddh $f2, $f2, $f0                            \r\n"
+        "psubh $f0, $f0, $f8                            \r\n"
+        "mov.d $f8, $f6                                 \r\n"
+        "paddh $f6, $f6, $f2                            \r\n"
+        "psubh $f2, $f2, $f8                            \r\n"
+        "mov.d $f8, $f4                                 \r\n"
+        "paddh $f4, $f4, $f0                            \r\n"
+        "psubh $f0, $f0, $f8                            \r\n"
+        "mov.d $f8, $f6                                 \r\n"
+        "punpcklhw $f6, $f6, $f2                        \r\n"
+        "punpckhhw $f8, $f8, $f2                        \r\n"
+        "punpckhhw $f2, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhwd $f4, $f6, $f0                        \r\n"
+        "punpcklwd $f6, $f6, $f0                        \r\n"
+        "mov.d $f0, $f8                                 \r\n"
+        "punpcklwd $f8, $f8, $f2                        \r\n"
+        "punpckhwd $f0, $f0, $f2                        \r\n"
+        "mov.d $f2, $f0                                 \r\n"
+        "paddh $f0, $f0, $f8                            \r\n"
+        "psubh $f8, $f8, $f2                            \r\n"
+        "mov.d $f2, $f4                                 \r\n"
+        "paddh $f4, $f4, $f6                            \r\n"
+        "psubh $f6, $f6, $f2                            \r\n"
+        "mov.d $f2, $f0                                 \r\n"
+        "paddh $f0, $f0, $f4                            \r\n"
+        "psubh $f4, $f4, $f2                            \r\n"
+        "mov.d $f2, $f8                                 \r\n"
+        "daddiu $10, %[qmul], -0x7fff                   \r\n"
+        "paddh $f8, $f8, $f6                            \r\n"
+        "bgtz $10, 1f                                   \r\n"
+        "psubh $f6, $f6, $f2                            \r\n"
+        "ori $10, $0, 0x80                              \r\n"
+        "dsll $10, $10, 0x10                            \r\n"
+        "punpckhhw $f2, $f0, %[ff_pw_1]                 \r\n"
+        "daddu %[qmul], %[qmul], $10                    \r\n"
+        "punpcklhw $f0, $f0, %[ff_pw_1]                 \r\n"
+        "punpckhhw $f10, $f4, %[ff_pw_1]                \r\n"
+        "punpcklhw $f4, $f4, %[ff_pw_1]                 \r\n"
+        "mtc1 %[qmul], $f14                             \r\n"
+        "punpcklwd $f14, $f14, $f14                     \r\n"
+        "pmaddhw $f0, $f0, $f14                         \r\n"
+        "pmaddhw $f4, $f4, $f14                         \r\n"
+        "pmaddhw $f2, $f2, $f14                         \r\n"
+        "pmaddhw $f10, $f10, $f14                       \r\n"
+        "psraw $f0, $f0, $f16                           \r\n"
+        "psraw $f4, $f4, $f16                           \r\n"
+        "psraw $f2, $f2, $f16                           \r\n"
+        "psraw $f10, $f10, $f16                         \r\n"
+        "packsswh $f0, $f0, $f2                         \r\n"
+        "packsswh $f4, $f4, $f10                        \r\n"
+        "mfc1 $9, $f0                                   \r\n"
+        "dsrl $f0, $f0, $f18                            \r\n"
+        "mfc1 %[input], $f0                             \r\n"
+        "sh $9, 0x0(%[output])                          \r\n"
+        "sh %[input], 0x80(%[output])                   \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x20(%[output])                         \r\n"
+        "sh %[input], 0xa0(%[output])                   \r\n"
+        "mfc1 $9, $f4                                   \r\n"
+        "dsrl $f4, $f4, $f18                            \r\n"
+        "mfc1 %[input], $f4                             \r\n"
+        "sh $9, 0x40(%[output])                         \r\n"
+        "sh %[input], 0xc0(%[output])                   \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x60(%[output])                         \r\n"
+        "sh %[input], 0xe0(%[output])                   \r\n"
+        "punpckhhw $f2, $f6, %[ff_pw_1]                 \r\n"
+        "punpcklhw $f6, $f6, %[ff_pw_1]                 \r\n"
+        "punpckhhw $f10, $f8, %[ff_pw_1]                \r\n"
+        "punpcklhw $f8, $f8, %[ff_pw_1]                 \r\n"
+        "mtc1 %[qmul], $f14                             \r\n"
+        "punpcklwd $f14, $f14, $f14                     \r\n"
+        "pmaddhw $f6, $f6, $f14                         \r\n"
+        "pmaddhw $f8, $f8, $f14                         \r\n"
+        "pmaddhw $f2, $f2, $f14                         \r\n"
+        "pmaddhw $f10, $f10, $f14                       \r\n"
+        "psraw $f6, $f6, $f16                           \r\n"
+        "psraw $f8, $f8, $f16                           \r\n"
+        "psraw $f2, $f2, $f16                           \r\n"
+        "psraw $f10, $f10, $f16                         \r\n"
+        "packsswh $f6, $f6, $f2                         \r\n"
+        "packsswh $f8, $f8, $f10                        \r\n"
+        "mfc1 $9, $f6                                   \r\n"
+        "dsrl $f6, $f6, $f18                            \r\n"
+        "mfc1 %[input], $f6                             \r\n"
+        "sh $9, 0x100(%[output])                        \r\n"
+        "sh %[input], 0x180(%[output])                  \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x120(%[output])                        \r\n"
+        "sh %[input], 0x1a0(%[output])                  \r\n"
+        "mfc1 $9, $f8                                   \r\n"
+        "dsrl $f8, $f8, $f18                            \r\n"
+        "mfc1 %[input], $f8                             \r\n"
+        "sh $9, 0x140(%[output])                        \r\n"
+        "sh %[input], 0x1c0(%[output])                  \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x160(%[output])                        \r\n"
+        "jr $31                                         \r\n"
+        "sh %[input], 0x1e0(%[output])                  \r\n"
+        "1:                                             \r\n"
+        "ori $10, $0, 0x1f                              \r\n"
+        "clz $9, %[qmul]                                \r\n"
+        "ori %[input], $0, 0x7                          \r\n"
+        "dsubu $9, $10, $9                              \r\n"
+        "ori $10, $0, 0x80                              \r\n"
+        "dsll $10, $10, 0x10                            \r\n"
+        "daddu %[qmul], %[qmul], $10                    \r\n"
+        "dsubu $10, $9, %[input]                        \r\n"
+        "movn $9, %[input], $10                         \r\n"
+        "daddiu %[input], %[input], 0x1                 \r\n"
+        "andi $10, $9, 0xff                             \r\n"
+        "dsrlv %[qmul], %[qmul], $10                    \r\n"
+        "dsubu %[input], %[input], $9                   \r\n"
+        "mtc1 %[input], $f12                            \r\n"
+        "punpckhhw $f2, $f0, %[ff_pw_1]                 \r\n"
+        "punpcklhw $f0, $f0, %[ff_pw_1]                 \r\n"
+        "punpckhhw $f10, $f4, %[ff_pw_1]                \r\n"
+        "punpcklhw $f4, $f4, %[ff_pw_1]                 \r\n"
+        "mtc1 %[qmul], $f14                             \r\n"
+        "punpcklwd $f14, $f14, $f14                     \r\n"
+        "pmaddhw $f0, $f0, $f14                         \r\n"
+        "pmaddhw $f4, $f4, $f14                         \r\n"
+        "pmaddhw $f2, $f2, $f14                         \r\n"
+        "pmaddhw $f10, $f10, $f14                       \r\n"
+        "psraw $f0, $f0, $f12                           \r\n"
+        "psraw $f4, $f4, $f12                           \r\n"
+        "psraw $f2, $f2, $f12                           \r\n"
+        "psraw $f10, $f10, $f12                         \r\n"
+        "packsswh $f0, $f0, $f2                         \r\n"
+        "packsswh $f4, $f4, $f10                        \r\n"
+        "mfc1 $9, $f0                                   \r\n"
+        "dsrl $f0, $f0, $f18                            \r\n"
+        "sh $9, 0x0(%[output])                          \r\n"
+        "mfc1 %[input], $f0                             \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "sh %[input], 0x80(%[output])                   \r\n"
+        "sh $9, 0x20(%[output])                         \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "mfc1 $9, $f4                                   \r\n"
+        "sh %[input], 0xa0(%[output])                   \r\n"
+        "dsrl $f4, $f4, $f18                            \r\n"
+        "sh $9, 0x40(%[output])                         \r\n"
+        "mfc1 %[input], $f4                             \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "sh %[input], 0xc0(%[output])                   \r\n"
+        "sh $9, 0x60(%[output])                         \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh %[input], 0xe0(%[output])                   \r\n"
+        "punpckhhw $f2, $f6, %[ff_pw_1]                 \r\n"
+        "punpcklhw $f6, $f6, %[ff_pw_1]                 \r\n"
+        "punpckhhw $f10, $f8, %[ff_pw_1]                \r\n"
+        "punpcklhw $f8, $f8, %[ff_pw_1]                 \r\n"
+        "mtc1 %[qmul], $f14                             \r\n"
+        "punpcklwd $f14, $f14, $f14                     \r\n"
+        "pmaddhw $f6, $f6, $f14                         \r\n"
+        "pmaddhw $f8, $f8, $f14                         \r\n"
+        "pmaddhw $f2, $f2, $f14                         \r\n"
+        "pmaddhw $f10, $f10, $f14                       \r\n"
+        "psraw $f6, $f6, $f12                           \r\n"
+        "psraw $f8, $f8, $f12                           \r\n"
+        "psraw $f2, $f2, $f12                           \r\n"
+        "psraw $f10, $f10, $f12                         \r\n"
+        "packsswh $f6, $f6, $f2                         \r\n"
+        "packsswh $f8, $f8, $f10                        \r\n"
+        "mfc1 $9, $f6                                   \r\n"
+        "dsrl $f6, $f6, $f18                            \r\n"
+        "mfc1 %[input], $f6                             \r\n"
+        "sh $9, 0x100(%[output])                        \r\n"
+        "sh %[input], 0x180(%[output])                  \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x120(%[output])                        \r\n"
+        "sh %[input], 0x1a0(%[output])                  \r\n"
+        "mfc1 $9, $f8                                   \r\n"
+        "dsrl $f8, $f8, $f18                            \r\n"
+        "mfc1 %[input], $f8                             \r\n"
+        "sh $9, 0x140(%[output])                        \r\n"
+        "sh %[input], 0x1c0(%[output])                  \r\n"
+        "dsrl $9, $9, 0x10                              \r\n"
+        "dsrl %[input], %[input], 0x10                  \r\n"
+        "sh $9, 0x160(%[output])                        \r\n"
+        "sh %[input], 0x1e0(%[output])                  \r\n"
+        ".set reorder                                   \r\n"
+        ::[output]"r"(output),[input]"r"(input),[qmul]"r"((uint64_t)qmul),
+          [ff_pw_1]"f"(ff_pw_1)
+        : "$9","$10","$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14","$f16",
+          "$f18"
+    );
+}
+
+void ff_h264_chroma422_dc_dequant_idct_8_mmi(int16_t *block, int qmul)
+{
+    int temp[8];
+    int t[8];
+
+    temp[0] = block[0] + block[16];
+    temp[1] = block[0] - block[16];
+    temp[2] = block[32] + block[48];
+    temp[3] = block[32] - block[48];
+    temp[4] = block[64] + block[80];
+    temp[5] = block[64] - block[80];
+    temp[6] = block[96] + block[112];
+    temp[7] = block[96] - block[112];
+
+    t[0] = temp[0] + temp[4] + temp[2] + temp[6];
+    t[1] = temp[0] - temp[4] + temp[2] - temp[6];
+    t[2] = temp[0] - temp[4] - temp[2] + temp[6];
+    t[3] = temp[0] + temp[4] - temp[2] - temp[6];
+    t[4] = temp[1] + temp[5] + temp[3] + temp[7];
+    t[5] = temp[1] - temp[5] + temp[3] - temp[7];
+    t[6] = temp[1] - temp[5] - temp[3] + temp[7];
+    t[7] = temp[1] + temp[5] - temp[3] - temp[7];
+
+    block[  0]= (t[0]*qmul + 128) >> 8;
+    block[ 32]= (t[1]*qmul + 128) >> 8;
+    block[ 64]= (t[2]*qmul + 128) >> 8;
+    block[ 96]= (t[3]*qmul + 128) >> 8;
+    block[ 16]= (t[4]*qmul + 128) >> 8;
+    block[ 48]= (t[5]*qmul + 128) >> 8;
+    block[ 80]= (t[6]*qmul + 128) >> 8;
+    block[112]= (t[7]*qmul + 128) >> 8;
+}
+
+void ff_h264_chroma_dc_dequant_idct_8_mmi(int16_t *block, int qmul)
+{
+    int a,b,c,d;
+
+    d = block[0] - block[16];
+    a = block[0] + block[16];
+    b = block[32] - block[48];
+    c = block[32] + block[48];
+    block[0] = ((a+c)*qmul) >> 7;
+    block[16]= ((d+b)*qmul) >> 7;
+    block[32]= ((a-c)*qmul) >> 7;
+    block[48]= ((d-b)*qmul) >> 7;
+}
+
 void ff_h264_weight_pixels16_8_mmi(uint8_t *block, int stride,
         int height, int log2_denom, int weight, int offset)
 {
@@ -276,3 +1282,1219 @@ void ff_h264_biweight_pixels4_8_mmi(uint8_t *dst, uint8_t *src,
         );
     }
 }
+
+static void inline chroma_inter_body_mmi(uint8_t *pix, int stride,
+        int alpha, int beta, int8_t *tc0)
+{
+    __asm__ volatile (
+        "xor $f16, $f16, $f16                           \r\n"
+        "mtc1 %[alpha], $f8                             \r\n"
+        "mtc1 %[beta], $f10                             \r\n"
+        "pshufh $f8, $f8, $f16                          \r\n"
+        "pshufh $f10, $f10, $f16                        \r\n"
+        "packushb $f8, $f8, $f8                         \r\n"
+        "packushb $f10, $f10, $f10                      \r\n"
+        "psubusb $f12, $f4, $f2                         \r\n"
+        "psubusb $f14, $f2, $f4                         \r\n"
+        "or $f14, $f14, $f12                            \r\n"
+        "psubusb $f14, $f14, $f8                        \r\n"
+        "psubusb $f12, $f2, $f0                         \r\n"
+        "psubusb $f8, $f0, $f2                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "psubusb $f12, $f4, $f6                         \r\n"
+        "psubusb $f8, $f6, $f4                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "xor $f12, $f12, $f12                           \r\n"
+        "pcmpeqb $f14, $f14, $f12                       \r\n"
+        "lwc1 $f12, 0x0(%[tc0])                         \r\n"
+        "punpcklbh $f12, $f12, $f12                     \r\n"
+        "and $f14, $f14, $f12                           \r\n"
+        "pcmpeqb $f8, $f8, $f8                          \r\n"
+        "xor $f10, $f2, $f4                             \r\n"
+        "xor $f6, $f6, $f8                              \r\n"
+        "and $f10, $f10, %[ff_pb_1]                     \r\n"
+        "pavgb $f6, $f6, $f0                            \r\n"
+        "xor $f8, $f8, $f2                              \r\n"
+        "pavgb $f6, $f6, %[ff_pb_3]                     \r\n"
+        "pavgb $f8, $f8, $f4                            \r\n"
+        "pavgb $f6, $f6, $f10                           \r\n"
+        "paddusb $f6, $f6, $f8                          \r\n"
+        "psubusb $f12, %[ff_pb_A1], $f6                 \r\n"
+        "psubusb $f6, $f6, %[ff_pb_A1]                  \r\n"
+        "pminub $f12, $f12, $f14                        \r\n"
+        "pminub $f6, $f6, $f14                          \r\n"
+        "psubusb $f2, $f2, $f12                         \r\n"
+        "psubusb $f4, $f4, $f6                          \r\n"
+        "paddusb $f2, $f2, $f6                          \r\n"
+        "paddusb $f4, $f4, $f12                         \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),
+          [alpha]"r"((int64_t)alpha),[beta]"r"((int64_t)beta),[tc0]"r"(tc0),
+          [ff_pb_1]"f"(ff_pb_1),[ff_pb_3]"f"(ff_pb_3),[ff_pb_A1]"f"(ff_pb_A1)
+        : "$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14","$f16"
+    );
+}
+
+static void inline chroma_intra_body_mmi(uint8_t *pix, int stride,
+        int alpha, int beta)
+{
+    __asm__ volatile (
+        "xor $f16, $f16, $f16                           \r\n"
+        "mtc1 %[alpha], $f8                             \r\n"
+        "mtc1 %[beta], $f10                             \r\n"
+        "pshufh $f8, $f8, $f16                          \r\n"
+        "pshufh $f10, $f10, $f16                        \r\n"
+        "packushb $f8, $f8, $f8                         \r\n"
+        "packushb $f10, $f10, $f10                      \r\n"
+        "psubusb $f12, $f4, $f2                         \r\n"
+        "psubusb $f14, $f2, $f4                         \r\n"
+        "or $f14, $f14, $f12                            \r\n"
+        "psubusb $f14, $f14, $f8                        \r\n"
+        "psubusb $f12, $f2, $f0                         \r\n"
+        "psubusb $f8, $f0, $f2                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "psubusb $f12, $f4, $f6                         \r\n"
+        "psubusb $f8, $f6, $f4                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "xor $f12, $f12, $f12                           \r\n"
+        "pcmpeqb $f14, $f14, $f12                       \r\n"
+        "mov.d $f10, $f2                                \r\n"
+        "mov.d $f12, $f4                                \r\n"
+        "xor $f8, $f2, $f6                              \r\n"
+        "and $f8, $f8, %[ff_pb_1]                       \r\n"
+        "pavgb $f2, $f2, $f6                            \r\n"
+        "psubusb $f2, $f2, $f8                          \r\n"
+        "pavgb $f2, $f2, $f0                            \r\n"
+        "xor $f8, $f4, $f0                              \r\n"
+        "and $f8, $f8, %[ff_pb_1]                       \r\n"
+        "pavgb $f4, $f4, $f0                            \r\n"
+        "psubusb $f4, $f4, $f8                          \r\n"
+        "pavgb $f4, $f4, $f6                            \r\n"
+        "psubb $f2, $f2, $f10                           \r\n"
+        "psubb $f4, $f4, $f12                           \r\n"
+        "and $f2, $f2, $f14                             \r\n"
+        "and $f4, $f4, $f14                             \r\n"
+        "paddb $f2, $f2, $f10                           \r\n"
+        "paddb $f4, $f4, $f12                           \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),
+          [alpha]"r"((int64_t)alpha),[beta]"r"((int64_t)beta),
+          [ff_pb_1]"f"(ff_pb_1)
+        : "$f0","$f2","$f4","$f8","$f10","$f12","$f14","$f16"
+    );
+}
+
+void ff_deblock_v8_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0)
+{
+    __asm__ volatile (
+        "daddu $8, %[stride], %[stride]                 \r\n"
+        "xor $f16, $f16, $f16                           \r\n"
+        "daddu $9, %[stride], $8                        \r\n"
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "dsubu $9, $0, $9                               \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "daddu $9, $9, %[pix]                           \r\n"
+        "ldc1 $f4, 0x0(%[pix])                          \r\n"
+        "gsldxc1 $f0, 0x0($9, %[stride])                \r\n"
+        "gsldxc1 $f2, 0x0($9, $8)                       \r\n"
+        "gsldxc1 $f6, 0x0(%[pix], %[stride])            \r\n"
+        "mtc1 %[alpha], $f8                             \r\n"
+        "mtc1 %[beta], $f10                             \r\n"
+        "pshufh $f8, $f8, $f16                          \r\n"
+        "pshufh $f10, $f10, $f16                        \r\n"
+        "packushb $f8, $f8, $f8                         \r\n"
+        "packushb $f10, $f10, $f10                      \r\n"
+        "psubusb $f12, $f4, $f2                         \r\n"
+        "psubusb $f14, $f2, $f4                         \r\n"
+        "or $f14, $f14, $f12                            \r\n"
+        "psubusb $f12, $f2, $f0                         \r\n"
+        "psubusb $f14, $f14, $f8                        \r\n"
+        "psubusb $f8, $f0, $f2                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f12, $f4, $f6                         \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "psubusb $f8, $f6, $f4                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "pcmpeqb $f14, $f14, $f16                       \r\n"
+        "pcmpeqb $f6, $f6, $f6                          \r\n"
+        "gslwlc1 $f8, 0x3(%[tc0])                       \r\n"
+        "gslwrc1 $f8, 0x0(%[tc0])                       \r\n"
+        "punpcklbh $f8, $f8, $f8                        \r\n"
+        "punpcklbh $f18, $f8, $f8                       \r\n"
+        "pcmpgtb $f8, $f18, $f6                         \r\n"
+        "ldc1 $f6, 0x0($9)                              \r\n"
+        "and $f20, $f8, $f14                            \r\n"
+        "psubusb $f14, $f6, $f2                         \r\n"
+        "psubusb $f12, $f2, $f6                         \r\n"
+        "psubusb $f14, $f14, $f10                       \r\n"
+        "psubusb $f12, $f12, $f10                       \r\n"
+        "pcmpeqb $f12, $f12, $f14                       \r\n"
+        "and $f12, $f12, $f20                           \r\n"
+        "and $f8, $f20, $f18                            \r\n"
+        "psubb $f14, $f8, $f12                          \r\n"
+        "and $f12, $f12, $f8                            \r\n"
+        "pavgb $f8, $f2, $f4                            \r\n"
+        "ldc1 $f22, 0x0($9)                             \r\n"
+        "pavgb $f6, $f6, $f8                            \r\n"
+        "xor $f8, $f8, $f22                             \r\n"
+        "and $f8, $f8, %[ff_pb_1]                       \r\n"
+        "psubusb $f6, $f6, $f8                          \r\n"
+        "psubusb $f8, $f0, $f12                         \r\n"
+        "paddusb $f12, $f12, $f0                        \r\n"
+        "pmaxub $f6, $f6, $f8                           \r\n"
+        "pminub $f6, $f6, $f12                          \r\n"
+        "gssdxc1 $f6, 0x0($9, %[stride])                \r\n"
+        "gsldxc1 $f8, 0x0(%[pix], $8)                   \r\n"
+        "psubusb $f6, $f8, $f4                          \r\n"
+        "psubusb $f12, $f4, $f8                         \r\n"
+        "psubusb $f6, $f6, $f10                         \r\n"
+        "psubusb $f12, $f12, $f10                       \r\n"
+        "pcmpeqb $f12, $f12, $f6                        \r\n"
+        "and $f12, $f12, $f20                           \r\n"
+        "psubb $f14, $f14, $f12                         \r\n"
+        "and $f10, $f18, $f12                           \r\n"
+        "gsldxc1 $f6, 0x0(%[pix], %[stride])            \r\n"
+        "pavgb $f12, $f2, $f4                           \r\n"
+        "gsldxc1 $f22, 0x0(%[pix], $8)                  \r\n"
+        "pavgb $f8, $f8, $f12                           \r\n"
+        "xor $f12, $f12, $f22                           \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "psubusb $f8, $f8, $f12                         \r\n"
+        "psubusb $f12, $f6, $f10                        \r\n"
+        "paddusb $f10, $f10, $f6                        \r\n"
+        "pmaxub $f8, $f8, $f12                          \r\n"
+        "pminub $f8, $f8, $f10                          \r\n"
+        "gssdxc1 $f8, 0x0(%[pix], %[stride])            \r\n"
+        "xor $f10, $f2, $f4                             \r\n"
+        "pcmpeqb $f8, $f8, $f8                          \r\n"
+        "and $f10, $f10, %[ff_pb_1]                     \r\n"
+        "xor $f6, $f6, $f8                              \r\n"
+        "xor $f8, $f8, $f2                              \r\n"
+        "pavgb $f6, $f6, $f0                            \r\n"
+        "pavgb $f6, $f6, %[ff_pb_3]                     \r\n"
+        "pavgb $f8, $f8, $f4                            \r\n"
+        "pavgb $f6, $f6, $f10                           \r\n"
+        "paddusb $f6, $f6, $f8                          \r\n"
+        "psubusb $f12, %[ff_pb_A1], $f6                 \r\n"
+        "psubusb $f6, $f6, %[ff_pb_A1]                  \r\n"
+        "pminub $f12, $f12, $f14                        \r\n"
+        "pminub $f6, $f6, $f14                          \r\n"
+        "psubusb $f2, $f2, $f12                         \r\n"
+        "psubusb $f4, $f4, $f6                          \r\n"
+        "paddusb $f2, $f2, $f6                          \r\n"
+        "paddusb $f4, $f4, $f12                         \r\n"
+        "gssdxc1 $f2, 0x0($9, $8)                       \r\n"
+        "sdc1 $f4, 0x0(%[pix])                          \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),
+          [alpha]"r"((int64_t)alpha),[beta]"r"((int64_t)beta),[tc0]"r"(tc0),
+          [ff_pb_1]"f"(ff_pb_1),[ff_pb_3]"f"(ff_pb_3),[ff_pb_A1]"f"(ff_pb_A1)
+        : "$8","$9","$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14","$f16",
+          "$f18","$f20","$f22"
+    );
+}
+
+void ff_deblock_v8_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta)
+{
+    uint64_t stack[0xa];
+
+    __asm__ volatile (
+        "ori $8, $0, 0x1                                \r\n"
+        "xor $f30, $f30, $f30                           \r\n"
+        "dmtc1 $8, $f16                                 \r\n"
+        "dsll $8, %[stride], 2                          \r\n"
+        "daddu $10, %[stride], %[stride]                \r\n"
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "dsll $f20, $f16, $f16                          \r\n"
+        "bltz %[alpha], 1f                              \r\n"
+        "daddu $9, $10, %[stride]                       \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "bltz %[beta], 1f                               \r\n"
+        "dsubu $8, $0, $8                               \r\n"
+        "daddu $8, $8, %[pix]                           \r\n"
+        "ldc1 $f4, 0x0(%[pix])                          \r\n"
+        "gsldxc1 $f0, 0x0($8, $10)                      \r\n"
+        "gsldxc1 $f2, 0x0($8, $9)                       \r\n"
+        "gsldxc1 $f6, 0x0(%[pix], %[stride])            \r\n"
+        "mtc1 %[alpha], $f8                             \r\n"
+        "mtc1 %[beta], $f10                             \r\n"
+        "pshufh $f8, $f8, $f30                          \r\n"
+        "pshufh $f10, $f10, $f30                        \r\n"
+        "packushb $f8, $f8, $f8                         \r\n"
+        "psubusb $f12, $f4, $f2                         \r\n"
+        "psubusb $f14, $f2, $f4                         \r\n"
+        "packushb $f10, $f10, $f10                      \r\n"
+        "or $f14, $f14, $f12                            \r\n"
+        "sdc1 $f8, 0x10+%[stack]                        \r\n"
+        "psubusb $f14, $f14, $f8                        \r\n"
+        "psubusb $f12, $f2, $f0                         \r\n"
+        "psubusb $f8, $f0, $f2                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "psubusb $f12, $f4, $f6                         \r\n"
+        "psubusb $f8, $f6, $f4                          \r\n"
+        "or $f8, $f8, $f12                              \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "or $f14, $f14, $f8                             \r\n"
+        "xor $f12, $f12, $f12                           \r\n"
+        "ldc1 $f8, 0x10+%[stack]                        \r\n"
+        "pcmpeqb $f14, $f14, $f12                       \r\n"
+        "sdc1 $f14, 0x20+%[stack]                       \r\n"
+        "pavgb $f8, $f8, $f30                           \r\n"
+        "psubusb $f14, $f4, $f2                         \r\n"
+        "pavgb $f8, $f8, %[ff_pb_1]                     \r\n"
+        "psubusb $f12, $f2, $f4                         \r\n"
+        "psubusb $f14, $f14, $f8                        \r\n"
+        "psubusb $f12, $f12, $f8                        \r\n"
+        "ldc1 $f28, 0x20+%[stack]                       \r\n"
+        "pcmpeqb $f12, $f12, $f14                       \r\n"
+        "and $f12, $f12, $f28                           \r\n"
+        "gsldxc1 $f28, 0x0($8, %[stride])               \r\n"
+        "psubusb $f14, $f28, $f2                        \r\n"
+        "psubusb $f8, $f2, $f28                         \r\n"
+        "psubusb $f14, $f14, $f10                       \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "pcmpeqb $f8, $f8, $f14                         \r\n"
+        "and $f8, $f8, $f12                             \r\n"
+        "gsldxc1 $f26, 0x0(%[pix], $10)                 \r\n"
+        "sdc1 $f8, 0x30+%[stack]                        \r\n"
+        "psubusb $f14, $f26, $f4                        \r\n"
+        "psubusb $f8, $f4, $f26                         \r\n"
+        "psubusb $f14, $f14, $f10                       \r\n"
+        "psubusb $f8, $f8, $f10                         \r\n"
+        "pcmpeqb $f8, $f8, $f14                         \r\n"
+        "and $f8, $f8, $f12                             \r\n"
+        "sdc1 $f8, 0x40+%[stack]                        \r\n"
+        "pavgb $f8, $f28, $f0                           \r\n"
+        "pavgb $f10, $f2, $f4                           \r\n"
+        "pavgb $f8, $f8, $f10                           \r\n"
+        "sdc1 $f10, 0x10+%[stack]                       \r\n"
+        "paddb $f12, $f28, $f0                          \r\n"
+        "paddb $f14, $f2, $f4                           \r\n"
+        "paddb $f12, $f12, $f14                         \r\n"
+        "mov.d $f14, $f12                               \r\n"
+        "sdc1 $f12, 0x0+%[stack]                        \r\n"
+        "psrlh $f12, $f12, $f16                         \r\n"
+        "pavgb $f12, $f12, $f30                         \r\n"
+        "xor $f12, $f12, $f8                            \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "psubb $f8, $f8, $f12                           \r\n"
+        "pavgb $f10, $f28, $f6                          \r\n"
+        "psubb $f12, $f28, $f6                          \r\n"
+        "paddb $f14, $f14, $f14                         \r\n"
+        "psubb $f14, $f14, $f12                         \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "psubb $f10, $f10, $f12                         \r\n"
+        "ldc1 $f24, 0x10+%[stack]                       \r\n"
+        "pavgb $f10, $f10, $f0                          \r\n"
+        "psrlh $f14, $f14, $f20                         \r\n"
+        "pavgb $f10, $f10, $f24                         \r\n"
+        "pavgb $f14, $f14, $f30                         \r\n"
+        "xor $f14, $f14, $f10                           \r\n"
+        "and $f14, $f14, %[ff_pb_1]                     \r\n"
+        "psubb $f10, $f10, $f14                         \r\n"
+        "xor $f14, $f2, $f6                             \r\n"
+        "pavgb $f12, $f2, $f6                           \r\n"
+        "and $f14, $f14, %[ff_pb_1]                     \r\n"
+        "psubb $f12, $f12, $f14                         \r\n"
+        "ldc1 $f24, 0x30+%[stack]                       \r\n"
+        "pavgb $f12, $f12, $f0                          \r\n"
+        "ldc1 $f22, 0x20+%[stack]                       \r\n"
+        "xor $f10, $f10, $f12                           \r\n"
+        "xor $f12, $f12, $f2                            \r\n"
+        "and $f10, $f10, $f24                           \r\n"
+        "and $f12, $f12, $f22                           \r\n"
+        "xor $f10, $f10, $f12                           \r\n"
+        "xor $f10, $f10, $f2                            \r\n"
+        "gssdxc1 $f10, 0x0($8, $9)                      \r\n"
+        "ldc1 $f10, 0x0($8)                             \r\n"
+        "paddb $f12, $f28, $f10                         \r\n"
+        "pavgb $f10, $f10, $f28                         \r\n"
+        "ldc1 $f22, 0x0+%[stack]                        \r\n"
+        "pavgb $f10, $f10, $f8                          \r\n"
+        "paddb $f12, $f12, $f12                         \r\n"
+        "paddb $f12, $f12, $f22                         \r\n"
+        "psrlh $f12, $f12, $f20                         \r\n"
+        "pavgb $f12, $f12, $f30                         \r\n"
+        "xor $f12, $f12, $f10                           \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "ldc1 $f22, 0x30+%[stack]                       \r\n"
+        "psubb $f10, $f10, $f12                         \r\n"
+        "xor $f8, $f8, $f0                              \r\n"
+        "xor $f10, $f10, $f28                           \r\n"
+        "and $f8, $f8, $f22                             \r\n"
+        "and $f10, $f10, $f22                           \r\n"
+        "xor $f8, $f8, $f0                              \r\n"
+        "xor $f10, $f10, $f28                           \r\n"
+        "gssdxc1 $f8, 0x0($8, $10)                      \r\n"
+        "gssdxc1 $f10, 0x0($8, %[stride])               \r\n"
+        "pavgb $f8, $f26, $f6                           \r\n"
+        "pavgb $f10, $f4, $f2                           \r\n"
+        "pavgb $f8, $f8, $f10                           \r\n"
+        "sdc1 $f10, 0x10+%[stack]                       \r\n"
+        "paddb $f12, $f26, $f6                          \r\n"
+        "paddb $f14, $f4, $f2                           \r\n"
+        "paddb $f12, $f12, $f14                         \r\n"
+        "mov.d $f14, $f12                               \r\n"
+        "sdc1 $f12, 0x0+%[stack]                        \r\n"
+        "psrlh $f12, $f12, $f16                         \r\n"
+        "pavgb $f12, $f12, $f30                         \r\n"
+        "xor $f12, $f12, $f8                            \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "psubb $f8, $f8, $f12                           \r\n"
+        "pavgb $f10, $f26, $f0                          \r\n"
+        "paddb $f14, $f14, $f14                         \r\n"
+        "psubb $f12, $f26, $f0                          \r\n"
+        "psubb $f14, $f14, $f12                         \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "psubb $f10, $f10, $f12                         \r\n"
+        "ldc1 $f22, 0x10+%[stack]                       \r\n"
+        "pavgb $f10, $f10, $f6                          \r\n"
+        "pavgb $f10, $f10, $f22                         \r\n"
+        "psrlh $f14, $f14, $f20                         \r\n"
+        "pavgb $f14, $f14, $f30                         \r\n"
+        "xor $f14, $f14, $f10                           \r\n"
+        "and $f14, $f14, %[ff_pb_1]                     \r\n"
+        "psubb $f10, $f10, $f14                         \r\n"
+        "xor $f14, $f4, $f0                             \r\n"
+        "pavgb $f12, $f4, $f0                           \r\n"
+        "and $f14, $f14, %[ff_pb_1]                     \r\n"
+        "ldc1 $f22, 0x40+%[stack]                       \r\n"
+        "psubb $f12, $f12, $f14                         \r\n"
+        "ldc1 $f24, 0x20+%[stack]                       \r\n"
+        "pavgb $f12, $f12, $f6                          \r\n"
+        "xor $f10, $f10, $f12                           \r\n"
+        "xor $f12, $f12, $f4                            \r\n"
+        "and $f10, $f10, $f22                           \r\n"
+        "and $f12, $f12, $f24                           \r\n"
+        "xor $f10, $f10, $f12                           \r\n"
+        "xor $f10, $f10, $f4                            \r\n"
+        "sdc1 $f10, 0x0(%[pix])                         \r\n"
+        "gsldxc1 $f10, 0x0(%[pix], $9)                  \r\n"
+        "paddb $f12, $f26, $f10                         \r\n"
+        "pavgb $f10, $f10, $f26                         \r\n"
+        "ldc1 $f22, 0x0+%[stack]                        \r\n"
+        "pavgb $f10, $f10, $f8                          \r\n"
+        "paddb $f12, $f12, $f12                         \r\n"
+        "paddb $f12, $f12, $f22                         \r\n"
+        "psrlh $f12, $f12, $f20                         \r\n"
+        "pavgb $f12, $f12, $f30                         \r\n"
+        "xor $f12, $f12, $f10                           \r\n"
+        "and $f12, $f12, %[ff_pb_1]                     \r\n"
+        "ldc1 $f22, 0x40+%[stack]                       \r\n"
+        "psubb $f10, $f10, $f12                         \r\n"
+        "xor $f8, $f8, $f6                              \r\n"
+        "xor $f10, $f10, $f26                           \r\n"
+        "and $f8, $f8, $f22                             \r\n"
+        "and $f10, $f10, $f22                           \r\n"
+        "xor $f8, $f8, $f6                              \r\n"
+        "xor $f10, $f10, $f26                           \r\n"
+        "gssdxc1 $f8, 0x0(%[pix], %[stride])            \r\n"
+        "gssdxc1 $f10, 0x0(%[pix], $10)                 \r\n"
+        "1:                                             \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),
+          [alpha]"r"((int64_t)alpha),[beta]"r"((int64_t)beta),
+          [stack]"m"(stack[0]),[ff_pb_1]"f"(ff_pb_1)
+        : "$8","$9","$10","$f0","$f2","$f4","$f6","$f8","$f10","$f12","$f14",
+          "$f16","$f18","$f20","$f22","$f24","$f26","$f28","$f30"
+    );
+}
+
+void ff_deblock_v_chroma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0)
+{
+    __asm__ volatile (
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "or $16, $0, %[pix]                             \r\n"
+        "dsubu $16, $16, %[stride]                      \r\n"
+        "dsubu $16, $16, %[stride]                      \r\n"
+        "ldc1 $f0, 0x0($16)                             \r\n"
+        "gsldxc1 $f2, 0x0($16, %[stride])               \r\n"
+        "ldc1 $f4, 0x0(%[pix])                          \r\n"
+        "gsldxc1 $f6, 0x0(%[pix], %[stride])            \r\n"
+        : [pix]"+r"(pix),[stride]"+r"(stride),[alpha]"+r"(alpha),
+          [beta]"+r"(beta)
+        : [tc0]"r"(tc0)
+        : "$16","$f2","$f4"
+    );
+
+    chroma_inter_body_mmi(pix, stride, alpha, beta, tc0);
+
+    __asm__ volatile (
+        "gssdxc1 $f2, 0x0($16, %[stride])               \r\n"
+        "sdc1 $f4, 0x0(%[pix])                          \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride)
+        : "$16","$f2","$f4"
+    );
+}
+
+void ff_deblock_v_chroma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta)
+{
+    __asm__ volatile (
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "or $16, $0, %[pix]                             \r\n"
+        "dsubu $16, $16, %[stride]                      \r\n"
+        "dsubu $16, $16, %[stride]                      \r\n"
+        "ldc1 $f0, 0x0($16)                             \r\n"
+        "gsldxc1 $f2, 0x0($16, %[stride])               \r\n"
+        "ldc1 $f4, 0x0(%[pix])                          \r\n"
+        "gsldxc1 $f6, 0x0(%[pix], %[stride])            \r\n"
+        : [pix]"+r"(pix),[stride]"+r"(stride),[alpha]"+r"(alpha),
+          [beta]"+r"(beta)
+        ::"$16","$f0","$f2","$f4","$f6"
+    );
+
+    chroma_intra_body_mmi(pix, stride, alpha, beta);
+
+    __asm__ volatile (
+        "gssdxc1 $f2, 0x0($16, %[stride])               \r\n"
+        "sdc1 $f4, 0x0(%[pix])                          \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride)
+        : "$16","$f2","$f4"
+    );
+}
+
+void ff_deblock_h_chroma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0)
+{
+    __asm__ volatile (
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "daddu $16, %[stride], %[stride]                \r\n"
+        "daddiu %[pix], %[pix], -0x2                    \r\n"
+        "daddu $17, $16, %[stride]                      \r\n"
+        "daddu $19, $16, $16                            \r\n"
+        "or $18, $0, %[pix]                             \r\n"
+        "daddu %[pix], %[pix], $17                      \r\n"
+        "gslwlc1 $f0, 0x3($18)                          \r\n"
+        "daddu $12, $18, %[stride]                      \r\n"
+        "gslwrc1 $f0, 0x0($18)                          \r\n"
+        "gslwlc1 $f4, 0x3($12)                          \r\n"
+        "daddu $13, $18, $16                            \r\n"
+        "gslwrc1 $f4, 0x0($12)                          \r\n"
+        "gslwlc1 $f2, 0x3($13)                          \r\n"
+        "gslwrc1 $f2, 0x0($13)                          \r\n"
+        "gslwlc1 $f6, 0x3(%[pix])                       \r\n"
+        "gslwrc1 $f6, 0x0(%[pix])                       \r\n"
+        "punpcklbh $f0, $f0, $f4                        \r\n"
+        "punpcklbh $f2, $f2, $f6                        \r\n"
+        "daddu $12, %[pix], %[stride]                   \r\n"
+        "punpckhhw $f4, $f0, $f2                        \r\n"
+        "punpcklhw $f0, $f0, $f2                        \r\n"
+        "gslwlc1 $f8, 0x3($12)                          \r\n"
+        "daddu $13, %[pix], $16                         \r\n"
+        "gslwrc1 $f8, 0x0($12)                          \r\n"
+        "gslwlc1 $f12, 0x3($13)                         \r\n"
+        "daddu $12, %[pix], $17                         \r\n"
+        "gslwrc1 $f12, 0x0($13)                         \r\n"
+        "gslwlc1 $f10, 0x3($12)                         \r\n"
+        "daddu $13, %[pix], $19                         \r\n"
+        "gslwrc1 $f10, 0x0($12)                         \r\n"
+        "gslwlc1 $f14, 0x3($13)                         \r\n"
+        "gslwrc1 $f14, 0x0($13)                         \r\n"
+        "punpcklbh $f8, $f8, $f12                       \r\n"
+        "punpcklbh $f10, $f10, $f14                     \r\n"
+        "mov.d $f12, $f8                                \r\n"
+        "punpcklhw $f8, $f8, $f10                       \r\n"
+        "punpckhhw $f12, $f12, $f10                     \r\n"
+        "punpckhwd $f2, $f0, $f8                        \r\n"
+        "punpckhwd $f6, $f4, $f12                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpcklwd $f4, $f4, $f12                       \r\n"
+        "mov.d $f20, $f0                                \r\n"
+        "mov.d $f22, $f6                                \r\n"
+        : [pix]"+r"(pix),[stride]"+r"(stride),[alpha]"+r"(alpha),
+          [beta]"+r"(beta)
+        ::"$12","$13","$16","$17","$18","$19","$f0","$f2","$f4","$f6","$f8",
+          "$f10","$f12","$f14","$f20","$f22"
+    );
+
+    chroma_inter_body_mmi(pix, stride, alpha, beta, tc0);
+
+    __asm__ volatile (
+        "punpckhwd $f8, $f20, $f20                      \r\n"
+        "punpckhwd $f10, $f2, $f2                       \r\n"
+        "punpckhwd $f12, $f4, $f4                       \r\n"
+        "punpcklbh $f0, $f20, $f2                       \r\n"
+        "punpcklbh $f4, $f4, $f22                       \r\n"
+        "punpcklhw $f2, $f0, $f4                        \r\n"
+        "punpckhhw $f0, $f0, $f4                        \r\n"
+        "gsswlc1 $f2, 0x3($18)                          \r\n"
+        "gsswrc1 $f2, 0x0($18)                          \r\n"
+        "daddu $12, $18, %[stride]                      \r\n"
+        "punpckhwd $f2, $f2, $f2                        \r\n"
+        "gsswlc1 $f2, 0x3($12)                          \r\n"
+        "daddu $13, $18, $16                            \r\n"
+        "gsswrc1 $f2, 0x0($12)                          \r\n"
+        "gsswlc1 $f0, 0x3($13)                          \r\n"
+        "gsswrc1 $f0, 0x0($13)                          \r\n"
+        "punpckhwd $f0, $f0, $f0                        \r\n"
+        "punpckhwd $f6, $f22, $f22                      \r\n"
+        "gsswlc1 $f0, 0x3(%[pix])                       \r\n"
+        "gsswrc1 $f0, 0x0(%[pix])                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "punpcklbh $f12, $f12, $f6                      \r\n"
+        "daddu $12, %[pix], %[stride]                   \r\n"
+        "punpcklhw $f10, $f8, $f12                      \r\n"
+        "punpckhhw $f8, $f8, $f12                       \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "punpckhwd $f10, $f10, $f10                     \r\n"
+        "daddu $12, %[pix], $16                         \r\n"
+        "daddu $13, %[pix], $17                         \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "gsswlc1 $f8, 0x3($13)                          \r\n"
+        "daddu $12, %[pix], $19                         \r\n"
+        "punpckhwd $f20, $f8, $f8                       \r\n"
+        "gsswrc1 $f8, 0x0($13)                          \r\n"
+        "gsswlc1 $f20, 0x3($12)                         \r\n"
+        "gsswrc1 $f20, 0x0($12)                         \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride)
+        : "$12","$13","$16","$17","$18","$19","$f0","$f2","$f4","$f6","$f8",
+          "$f10","$f12","$f20"
+    );
+}
+
+void ff_deblock_h_chroma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta)
+{
+    __asm__ volatile (
+        "daddiu %[alpha], %[alpha], -0x1                \r\n"
+        "daddiu %[beta], %[beta], -0x1                  \r\n"
+        "daddu $16, %[stride], %[stride]                \r\n"
+        "daddiu %[pix], %[pix], -0x2                    \r\n"
+        "daddu $17, $16, %[stride]                      \r\n"
+        "daddu $19, $16, $16                            \r\n"
+        "or $18, $0, %[pix]                             \r\n"
+        "daddu %[pix], %[pix], $17                      \r\n"
+        "gslwlc1 $f0, 0x3($18)                          \r\n"
+        "daddu $12, $18, %[stride]                      \r\n"
+        "gslwrc1 $f0, 0x0($18)                          \r\n"
+        "gslwlc1 $f4, 0x3($12)                          \r\n"
+        "daddu $13, $18, $16                            \r\n"
+        "gslwrc1 $f4, 0x0($12)                          \r\n"
+        "gslwlc1 $f2, 0x3($13)                          \r\n"
+        "gslwrc1 $f2, 0x0($13)                          \r\n"
+        "gslwlc1 $f6, 0x3(%[pix])                       \r\n"
+        "gslwrc1 $f6, 0x0(%[pix])                       \r\n"
+        "punpcklbh $f0, $f0, $f4                        \r\n"
+        "punpcklbh $f2, $f2, $f6                        \r\n"
+        "daddu $12, %[pix], %[stride]                   \r\n"
+        "punpckhhw $f4, $f0, $f2                        \r\n"
+        "punpcklhw $f0, $f0, $f2                        \r\n"
+        "gslwlc1 $f8, 0x3($12)                          \r\n"
+        "daddu $13, %[pix], $16                         \r\n"
+        "gslwrc1 $f8, 0x0($12)                          \r\n"
+        "gslwlc1 $f12, 0x3($13)                         \r\n"
+        "daddu $12, %[pix], $17                         \r\n"
+        "gslwrc1 $f12, 0x0($13)                         \r\n"
+        "gslwlc1 $f10, 0x3($12)                         \r\n"
+        "daddu $13, %[pix], $19                         \r\n"
+        "gslwrc1 $f10, 0x0($12)                         \r\n"
+        "gslwlc1 $f14, 0x3($13)                         \r\n"
+        "gslwrc1 $f14, 0x0($13)                         \r\n"
+        "punpcklbh $f8, $f8, $f12                       \r\n"
+        "punpcklbh $f10, $f10, $f14                     \r\n"
+        "mov.d $f12, $f8                                \r\n"
+        "punpcklhw $f8, $f8, $f10                       \r\n"
+        "punpckhhw $f12, $f12, $f10                     \r\n"
+        "punpckhwd $f2, $f0, $f8                        \r\n"
+        "punpckhwd $f6, $f4, $f12                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpcklwd $f4, $f4, $f12                       \r\n"
+        : [pix]"+r"(pix),[stride]"+r"(stride),[alpha]"+r"(alpha),
+          [beta]"+r"(beta)
+        ::"$12","$13","$16","$17","$18","$19","$f0","$f2","$f4","$f6","$f8",
+          "$f10","$f12","$f14","$f20","$f22"
+    );
+
+    chroma_intra_body_mmi(pix, stride, alpha, beta);
+
+    __asm__ volatile (
+        "punpckhwd $f8, $f0, $f0                        \r\n"
+        "punpckhwd $f10, $f2, $f2                       \r\n"
+        "punpckhwd $f12, $f4, $f4                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpcklhw $f2, $f0, $f4                        \r\n"
+        "punpckhhw $f0, $f0, $f4                        \r\n"
+        "gsswlc1 $f2, 0x3($18)                          \r\n"
+        "gsswrc1 $f2, 0x0($18)                          \r\n"
+        "daddu $12, $18, %[stride]                      \r\n"
+        "punpckhwd $f2, $f2, $f2                        \r\n"
+        "gsswlc1 $f2, 0x3($12)                          \r\n"
+        "daddu $13, $18, $16                            \r\n"
+        "gsswrc1 $f2, 0x0($12)                          \r\n"
+        "gsswlc1 $f0, 0x3($13)                          \r\n"
+        "gsswrc1 $f0, 0x0($13)                          \r\n"
+        "punpckhwd $f0, $f0, $f0                        \r\n"
+        "punpckhwd $f6, $f6, $f6                        \r\n"
+        "gsswlc1 $f0, 0x3(%[pix])                       \r\n"
+        "gsswrc1 $f0, 0x0(%[pix])                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "punpcklbh $f12, $f12, $f6                      \r\n"
+        "daddu $12, %[pix], %[stride]                   \r\n"
+        "punpcklhw $f10, $f8, $f12                      \r\n"
+        "punpckhhw $f8, $f8, $f12                       \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "punpckhwd $f10, $f10, $f10                     \r\n"
+        "daddu $12, %[pix], $16                         \r\n"
+        "daddu $13, %[pix], $17                         \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "gsswlc1 $f8, 0x3($13)                          \r\n"
+        "daddu $12, %[pix], $19                         \r\n"
+        "punpckhwd $f20, $f8, $f8                       \r\n"
+        "gsswrc1 $f8, 0x0($13)                          \r\n"
+        "gsswlc1 $f20, 0x3($12)                         \r\n"
+        "gsswrc1 $f20, 0x0($12)                         \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride)
+        : "$12","$13","$16","$17","$18","$19","$f0","$f2","$f4","$f6","$f8",
+          "$f10","$f12","$f20"
+    );
+}
+
+void ff_deblock_v_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0)
+{
+    if ((tc0[0] & tc0[1]) >= 0)
+        ff_deblock_v8_luma_8_mmi(pix + 0, stride, alpha, beta, tc0);
+    if ((tc0[2] & tc0[3]) >= 0)
+        ff_deblock_v8_luma_8_mmi(pix + 8, stride, alpha, beta, tc0 + 2);
+}
+
+void ff_deblock_v_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta)
+{
+    ff_deblock_v8_luma_intra_8_mmi(pix + 0, stride, alpha, beta);
+    ff_deblock_v8_luma_intra_8_mmi(pix + 8, stride, alpha, beta);
+}
+
+void ff_deblock_h_luma_8_mmi(uint8_t *pix, int stride, int alpha, int beta,
+        int8_t *tc0)
+{
+    uint64_t stack[0xd];
+
+    __asm__ volatile (
+        "daddu $15, %[stride], %[stride]                \r\n"
+        "daddiu $8, %[pix], -0x4                        \r\n"
+        "daddu $9, %[stride], $15                       \r\n"
+        "gsldlc1 $f0, 0x7($8)                           \r\n"
+        "gsldrc1 $f0, 0x0($8)                           \r\n"
+        "daddu $12, $8, %[stride]                       \r\n"
+        "daddu $10, $8, $9                              \r\n"
+        "gsldlc1 $f2, 0x7($12)                          \r\n"
+        "daddu $11, $8, $15                             \r\n"
+        "gsldrc1 $f2, 0x0($12)                          \r\n"
+        "gsldlc1 $f4, 0x7($11)                          \r\n"
+        "gsldrc1 $f4, 0x0($11)                          \r\n"
+        "gsldlc1 $f6, 0x7($10)                          \r\n"
+        "daddu $12, $10, %[stride]                      \r\n"
+        "gsldrc1 $f6, 0x0($10)                          \r\n"
+        "gsldlc1 $f8, 0x7($12)                          \r\n"
+        "daddu $11, $10, $15                            \r\n"
+        "gsldrc1 $f8, 0x0($12)                          \r\n"
+        "gsldlc1 $f10, 0x7($11)                         \r\n"
+        "daddu $12, $10, $9                             \r\n"
+        "gsldrc1 $f10, 0x0($11)                         \r\n"
+        "gsldlc1 $f12, 0x7($12)                         \r\n"
+        "gsldrc1 $f12, 0x0($12)                         \r\n"
+        "daddu $14, $15, $15                            \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "daddu $12, $10, $14                            \r\n"
+        "sdc1 $f2, 0x10+%[stack]                        \r\n"
+        "gsldlc1 $f16, 0x7($12)                         \r\n"
+        "gsldrc1 $f16, 0x0($12)                         \r\n"
+        "daddu $13, $14, $14                            \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "punpckhhw $f2, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "ldc1 $f16, 0x10+%[stack]                       \r\n"
+        "punpckhwd $f0, $f0, $f8                        \r\n"
+        "sdc1 $f0, 0x0+%[stack]                         \r\n"
+        "punpckhhw $f12, $f14, $f16                     \r\n"
+        "punpcklhw $f14, $f14, $f16                     \r\n"
+        "punpckhhw $f0, $f6, $f10                       \r\n"
+        "punpcklhw $f6, $f6, $f10                       \r\n"
+        "punpcklwd $f12, $f12, $f0                      \r\n"
+        "punpckhwd $f10, $f14, $f6                      \r\n"
+        "punpcklwd $f14, $f14, $f6                      \r\n"
+        "punpckhwd $f6, $f2, $f4                        \r\n"
+        "punpcklwd $f2, $f2, $f4                        \r\n"
+        "sdc1 $f2, 0x10+%[stack]                        \r\n"
+        "sdc1 $f6, 0x20+%[stack]                        \r\n"
+        "sdc1 $f14, 0x30+%[stack]                       \r\n"
+        "sdc1 $f10, 0x40+%[stack]                       \r\n"
+        "sdc1 $f12, 0x50+%[stack]                       \r\n"
+        "daddu $8, $8, $13                              \r\n"
+        "daddu $10, $10, $13                            \r\n"
+        "gsldlc1 $f0, 0x7($8)                           \r\n"
+        "daddu $12, $8, %[stride]                       \r\n"
+        "gsldrc1 $f0, 0x0($8)                           \r\n"
+        "gsldlc1 $f2, 0x7($12)                          \r\n"
+        "daddu $11, $8, $15                             \r\n"
+        "gsldrc1 $f2, 0x0($12)                          \r\n"
+        "gsldlc1 $f4, 0x7($11)                          \r\n"
+        "gsldrc1 $f4, 0x0($11)                          \r\n"
+        "gsldlc1 $f6, 0x7($10)                          \r\n"
+        "daddu $12, $10, %[stride]                      \r\n"
+        "gsldrc1 $f6, 0x0($10)                          \r\n"
+        "gsldlc1 $f8, 0x7($12)                          \r\n"
+        "daddu $11, $10, $15                            \r\n"
+        "gsldrc1 $f8, 0x0($12)                          \r\n"
+        "gsldlc1 $f10, 0x7($11)                         \r\n"
+        "daddu $12, $10, $9                             \r\n"
+        "gsldrc1 $f10, 0x0($11)                         \r\n"
+        "gsldlc1 $f12, 0x7($12)                         \r\n"
+        "gsldrc1 $f12, 0x0($12)                         \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "daddu $12, $10, $14                            \r\n"
+        "sdc1 $f2, 0x18+%[stack]                        \r\n"
+        "gsldlc1 $f16, 0x7($12)                         \r\n"
+        "gsldrc1 $f16, 0x0($12)                         \r\n"
+        "punpckhhw $f2, $f0, $f4                        \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "punpckhwd $f0, $f0, $f8                        \r\n"
+        "ldc1 $f16, 0x18+%[stack]                       \r\n"
+        "sdc1 $f0, 0x8+%[stack]                         \r\n"
+        "punpckhhw $f12, $f14, $f16                     \r\n"
+        "punpcklhw $f14, $f14, $f16                     \r\n"
+        "punpckhhw $f0, $f6, $f10                       \r\n"
+        "punpcklhw $f6, $f6, $f10                       \r\n"
+        "punpckhwd $f10, $f14, $f6                      \r\n"
+        "punpcklwd $f14, $f14, $f6                      \r\n"
+        "punpckhwd $f6, $f2, $f4                        \r\n"
+        "punpcklwd $f2, $f2, $f4                        \r\n"
+        "punpcklwd $f12, $f12, $f0                      \r\n"
+        "sdc1 $f2, 0x18+%[stack]                        \r\n"
+        "sdc1 $f6, 0x28+%[stack]                        \r\n"
+        "sdc1 $f14, 0x38+%[stack]                       \r\n"
+        "sdc1 $f10, 0x48+%[stack]                       \r\n"
+        "sdc1 $f12, 0x58+%[stack]                       \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),[stack]"m"(stack[0])
+        : "$8","$9","$10","$11","$12","$13","$14","$15","$f0","$f2","$f4",
+          "$f6","$f8","$f10","$f12","$f14","$f16"
+    );
+
+    ff_deblock_v_luma_8_mmi((uint8_t *) &stack[6], 0x10, alpha, beta, tc0);
+
+    __asm__ volatile (
+        "daddu $15, %[stride], %[stride]                \r\n"
+        "daddiu $8, %[pix], -0x2                        \r\n"
+        "daddu $14, $15, $15                            \r\n"
+        "daddu $9, $15, %[stride]                       \r\n"
+        "daddu $13, $14, $14                            \r\n"
+        "daddu $10, $8, $9                              \r\n"
+        "ldc1 $f0, 0x10+%[stack]                        \r\n"
+        "ldc1 $f2, 0x20+%[stack]                        \r\n"
+        "ldc1 $f4, 0x30+%[stack]                        \r\n"
+        "ldc1 $f6, 0x40+%[stack]                        \r\n"
+        "punpckhwd $f8, $f0, $f0                        \r\n"
+        "punpckhwd $f10, $f2, $f2                       \r\n"
+        "punpckhwd $f12, $f4, $f4                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpcklhw $f2, $f0, $f4                        \r\n"
+        "punpckhhw $f0, $f0, $f4                        \r\n"
+        "gsswlc1 $f2, 0x3($8)                           \r\n"
+        "gsswrc1 $f2, 0x0($8)                           \r\n"
+        "daddu $12, $8, %[stride]                       \r\n"
+        "punpckhwd $f2, $f2, $f2                        \r\n"
+        "daddu $11, $8, $15                             \r\n"
+        "gsswlc1 $f2, 0x3($12)                          \r\n"
+        "gsswrc1 $f2, 0x0($12)                          \r\n"
+        "gsswlc1 $f0, 0x3($11)                          \r\n"
+        "gsswrc1 $f0, 0x0($11)                          \r\n"
+        "punpckhwd $f0, $f0, $f0                        \r\n"
+        "punpckhwd $f6, $f6, $f6                        \r\n"
+        "gsswlc1 $f0, 0x3($10)                          \r\n"
+        "gsswrc1 $f0, 0x0($10)                          \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "punpcklbh $f12, $f12, $f6                      \r\n"
+        "punpcklhw $f10, $f8, $f12                      \r\n"
+        "daddu $12, $10, %[stride]                      \r\n"
+        "punpckhhw $f8, $f8, $f12                       \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "daddu $12, $10, $15                            \r\n"
+        "punpckhwd $f10, $f10, $f10                     \r\n"
+        "daddu $11, $10, $9                             \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "gsswlc1 $f8, 0x3($11)                          \r\n"
+        "gsswrc1 $f8, 0x0($11)                          \r\n"
+        "daddu $12, $10, $14                            \r\n"
+        "punpckhwd $f8, $f8, $f8                        \r\n"
+        "daddu $8, $8, $13                              \r\n"
+        "gsswlc1 $f8, 0x3($12)                          \r\n"
+        "gsswrc1 $f8, 0x0($12)                          \r\n"
+        "daddu $10, $10, $13                            \r\n"
+        "ldc1 $f0, 0x18+%[stack]                        \r\n"
+        "ldc1 $f2, 0x28+%[stack]                        \r\n"
+        "ldc1 $f4, 0x38+%[stack]                        \r\n"
+        "ldc1 $f6, 0x48+%[stack]                        \r\n"
+        "daddu $15, %[stride], %[stride]                \r\n"
+        "punpckhwd $f8, $f0, $f0                        \r\n"
+        "daddu $14, $15, $15                            \r\n"
+        "punpckhwd $f10, $f2, $f2                       \r\n"
+        "punpckhwd $f12, $f4, $f4                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "daddu $12, $8, %[stride]                       \r\n"
+        "punpcklhw $f2, $f0, $f4                        \r\n"
+        "punpckhhw $f0, $f0, $f4                        \r\n"
+        "gsswlc1 $f2, 0x3($8)                           \r\n"
+        "gsswrc1 $f2, 0x0($8)                           \r\n"
+        "punpckhwd $f2, $f2, $f2                        \r\n"
+        "daddu $11, $8, $15                             \r\n"
+        "gsswlc1 $f2, 0x3($12)                          \r\n"
+        "gsswrc1 $f2, 0x0($12)                          \r\n"
+        "gsswlc1 $f0, 0x3($11)                          \r\n"
+        "gsswrc1 $f0, 0x0($11)                          \r\n"
+        "punpckhwd $f0, $f0, $f0                        \r\n"
+        "punpckhwd $f6, $f6, $f6                        \r\n"
+        "gsswlc1 $f0, 0x3($10)                          \r\n"
+        "gsswrc1 $f0, 0x0($10)                          \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "punpcklbh $f12, $f12, $f6                      \r\n"
+        "daddu $12, $10, %[stride]                      \r\n"
+        "punpcklhw $f10, $f8, $f12                      \r\n"
+        "punpckhhw $f8, $f8, $f12                       \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "daddu $12, $10, $15                            \r\n"
+        "punpckhwd $f10, $f10, $f10                     \r\n"
+        "daddu $11, $10, $9                             \r\n"
+        "gsswlc1 $f10, 0x3($12)                         \r\n"
+        "gsswrc1 $f10, 0x0($12)                         \r\n"
+        "gsswlc1 $f8, 0x3($11)                          \r\n"
+        "gsswrc1 $f8, 0x0($11)                          \r\n"
+        "daddu $12, $10, $14                            \r\n"
+        "punpckhwd $f8, $f8, $f8                        \r\n"
+        "gsswlc1 $f8, 0x3($12)                          \r\n"
+        "gsswrc1 $f8, 0x0($12)                          \r\n"
+        ::[pix]"r"(pix),[stride]"r"((int64_t)stride),[stack]"m"(stack[0])
+        : "$8","$9","$10","$11","$12","$13","$14","$15","$f0","$f2","$f4",
+          "$f6","$f8","$f10","$f12","$f14","$f16"
+    );
+}
+
+void ff_deblock_h_luma_intra_8_mmi(uint8_t *pix, int stride, int alpha,
+        int beta)
+{
+    uint64_t ptmp[0x11];
+    uint64_t pdat[4];
+
+    __asm__ volatile (
+        "daddu $12, %[stride], %[stride]                \r\n"
+        "daddiu $10, %[pix], -0x4                       \r\n"
+        "daddu $11, $12, %[stride]                      \r\n"
+        "daddu $13, $12, $12                            \r\n"
+        "daddu $9, $10, $11                             \r\n"
+        "daddu $8, $10, %[stride]                       \r\n"
+        "gsldlc1 $f0, 0x7($10)                          \r\n"
+        "gsldrc1 $f0, 0x0($10)                          \r\n"
+        "daddu $14, $10, $12                            \r\n"
+        "gsldlc1 $f2, 0x7($8)                           \r\n"
+        "gsldrc1 $f2, 0x0($8)                           \r\n"
+        "gsldlc1 $f4, 0x7($14)                          \r\n"
+        "gsldrc1 $f4, 0x0($14)                          \r\n"
+        "daddu $8, $9, %[stride]                        \r\n"
+        "gsldlc1 $f6, 0x7($9)                           \r\n"
+        "gsldrc1 $f6, 0x0($9)                           \r\n"
+        "daddu $14, $9, $12                             \r\n"
+        "gsldlc1 $f8, 0x7($8)                           \r\n"
+        "gsldrc1 $f8, 0x0($8)                           \r\n"
+        "daddu $8, $9, $11                              \r\n"
+        "gsldlc1 $f10, 0x7($14)                         \r\n"
+        "gsldrc1 $f10, 0x0($14)                         \r\n"
+        "gsldlc1 $f12, 0x7($8)                          \r\n"
+        "gsldrc1 $f12, 0x0($8)                          \r\n"
+        "daddu $8, $9, $13                              \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "gsldlc1 $f16, 0x7($8)                          \r\n"
+        "gsldrc1 $f16, 0x0($8)                          \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "sdc1 $f6, 0x0+%[ptmp]                          \r\n"
+        "punpckhhw $f6, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "punpckhhw $f12, $f14, $f2                      \r\n"
+        "punpcklhw $f14, $f14, $f2                      \r\n"
+        "sdc1 $f4, 0x20+%[ptmp]                         \r\n"
+        "ldc1 $f4, 0x0+%[ptmp]                          \r\n"
+        "punpckhhw $f2, $f4, $f10                       \r\n"
+        "punpcklhw $f4, $f4, $f10                       \r\n"
+        "punpckhwd $f10, $f0, $f8                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpckhwd $f8, $f14, $f4                       \r\n"
+        "punpcklwd $f14, $f14, $f4                      \r\n"
+        "sdc1 $f0, 0x0+%[ptmp]                          \r\n"
+        "sdc1 $f10, 0x10+%[ptmp]                        \r\n"
+        "sdc1 $f14, 0x40+%[ptmp]                        \r\n"
+        "sdc1 $f8, 0x50+%[ptmp]                         \r\n"
+        "ldc1 $f16, 0x20+%[ptmp]                        \r\n"
+        "punpckhwd $f0, $f6, $f16                       \r\n"
+        "punpcklwd $f6, $f6, $f16                       \r\n"
+        "punpckhwd $f10, $f12, $f2                      \r\n"
+        "punpcklwd $f12, $f12, $f2                      \r\n"
+        "daddu $8, $13, $13                             \r\n"
+        "sdc1 $f6, 0x20+%[ptmp]                         \r\n"
+        "sdc1 $f0, 0x30+%[ptmp]                         \r\n"
+        "sdc1 $f12, 0x60+%[ptmp]                        \r\n"
+        "sdc1 $f10, 0x70+%[ptmp]                        \r\n"
+        "daddu $10, $10, $8                             \r\n"
+        "daddu $9, $9, $8                               \r\n"
+        "daddu $8, $10, %[stride]                       \r\n"
+        "gsldlc1 $f0, 0x7($10)                          \r\n"
+        "gsldrc1 $f0, 0x0($10)                          \r\n"
+        "daddu $14, $10, $12                            \r\n"
+        "gsldlc1 $f2, 0x7($8)                           \r\n"
+        "gsldrc1 $f2, 0x0($8)                           \r\n"
+        "gsldlc1 $f4, 0x7($14)                          \r\n"
+        "gsldrc1 $f4, 0x0($14)                          \r\n"
+        "daddu $8, $9, %[stride]                        \r\n"
+        "gsldlc1 $f6, 0x7($9)                           \r\n"
+        "gsldrc1 $f6, 0x0($9)                           \r\n"
+        "daddu $14, $9, $12                             \r\n"
+        "gsldlc1 $f8, 0x7($8)                           \r\n"
+        "gsldrc1 $f8, 0x0($8)                           \r\n"
+        "daddu $8, $9, $11                              \r\n"
+        "gsldlc1 $f10, 0x7($14)                         \r\n"
+        "gsldrc1 $f10, 0x0($14)                         \r\n"
+        "gsldlc1 $f12, 0x7($8)                          \r\n"
+        "gsldrc1 $f12, 0x0($8)                          \r\n"
+        "daddu $8, $9, $13                              \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "gsldlc1 $f16, 0x7($8)                          \r\n"
+        "gsldrc1 $f16, 0x0($8)                          \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "sdc1 $f6, 0x8+%[ptmp]                          \r\n"
+        "punpckhhw $f6, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "punpckhhw $f12, $f14, $f2                      \r\n"
+        "punpcklhw $f14, $f14, $f2                      \r\n"
+        "sdc1 $f4, 0x28+%[ptmp]                         \r\n"
+        "ldc1 $f4, 0x8+%[ptmp]                          \r\n"
+        "punpckhhw $f2, $f4, $f10                       \r\n"
+        "punpcklhw $f4, $f4, $f10                       \r\n"
+        "punpckhwd $f10, $f0, $f8                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpckhwd $f8, $f14, $f4                       \r\n"
+        "punpcklwd $f14, $f14, $f4                      \r\n"
+        "sdc1 $f0, 0x8+%[ptmp]                          \r\n"
+        "sdc1 $f10, 0x18+%[ptmp]                        \r\n"
+        "sdc1 $f14, 0x48+%[ptmp]                        \r\n"
+        "sdc1 $f8, 0x58+%[ptmp]                         \r\n"
+        "ldc1 $f16, 0x28+%[ptmp]                        \r\n"
+        "punpckhwd $f0, $f6, $f16                       \r\n"
+        "punpcklwd $f6, $f6, $f16                       \r\n"
+        "punpckhwd $f10, $f12, $f2                      \r\n"
+        "punpcklwd $f12, $f12, $f2                      \r\n"
+        "sdc1 $f6, 0x28+%[ptmp]                         \r\n"
+        "sdc1 $f0, 0x38+%[ptmp]                         \r\n"
+        "sdc1 $f12, 0x68+%[ptmp]                        \r\n"
+        "sdc1 $f10, 0x78+%[ptmp]                        \r\n"
+        "sd $10, 0x00+%[pdat]                           \r\n"
+        "sd $11, 0x08+%[pdat]                           \r\n"
+        "sd $12, 0x10+%[pdat]                           \r\n"
+        "sd $13, 0x18+%[pdat]                           \r\n"
+        ::[pix]"r"(pix),[stride]"r"((uint64_t)stride),[ptmp]"m"(ptmp[0]),
+          [pdat]"m"(pdat[0])
+        : "$8","$9","$10","$11","$12","$13","$14","$f0","$f2","$f4","$f6",
+          "$f8","$f10","$f12","$f14","$f16"
+    );
+
+    ff_deblock_v_luma_intra_8_mmi((uint8_t *) &ptmp[8], 0x10, alpha, beta);
+
+    __asm__ volatile (
+        "ld $10, 0x00+%[pdat]                           \r\n"
+        "ld $11, 0x08+%[pdat]                           \r\n"
+        "ld $12, 0x10+%[pdat]                           \r\n"
+        "ld $13, 0x18+%[pdat]                           \r\n"
+        "daddu $9, $10, $11                             \r\n"
+        "ldc1 $f0, 0x8+%[ptmp]                          \r\n"
+        "ldc1 $f2, 0x18+%[ptmp]                         \r\n"
+        "ldc1 $f4, 0x28+%[ptmp]                         \r\n"
+        "ldc1 $f6, 0x38+%[ptmp]                         \r\n"
+        "ldc1 $f8, 0x48+%[ptmp]                         \r\n"
+        "ldc1 $f10, 0x58+%[ptmp]                        \r\n"
+        "ldc1 $f12, 0x68+%[ptmp]                        \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "ldc1 $f16, 0x78+%[ptmp]                        \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "gssdlc1 $f6, 0x7($10)                          \r\n"
+        "gssdrc1 $f6, 0x0($10)                          \r\n"
+        "daddu $8, $10, $12                             \r\n"
+        "punpckhhw $f6, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "punpckhhw $f12, $f14, $f2                      \r\n"
+        "punpcklhw $f14, $f14, $f2                      \r\n"
+        "gssdlc1 $f4, 0x7($8)                           \r\n"
+        "gssdrc1 $f4, 0x0($8)                           \r\n"
+        "gsldlc1 $f4, 0x7($10)                          \r\n"
+        "gsldrc1 $f4, 0x0($10)                          \r\n"
+        "punpckhhw $f2, $f4, $f10                       \r\n"
+        "punpcklhw $f4, $f4, $f10                       \r\n"
+        "punpckhwd $f10, $f0, $f8                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpckhwd $f8, $f14, $f4                       \r\n"
+        "punpcklwd $f14, $f14, $f4                      \r\n"
+        "daddu $8, $10, %[stride]                       \r\n"
+        "gssdlc1 $f0, 0x7($10)                          \r\n"
+        "gssdrc1 $f0, 0x0($10)                          \r\n"
+        "daddu $14, $9, %[stride]                       \r\n"
+        "gssdlc1 $f10, 0x7($8)                          \r\n"
+        "gssdrc1 $f10, 0x0($8)                          \r\n"
+        "daddu $8, $9, $12                              \r\n"
+        "gssdlc1 $f14, 0x7($14)                         \r\n"
+        "gssdrc1 $f14, 0x0($14)                         \r\n"
+        "daddu $14, $10, $12                            \r\n"
+        "gssdlc1 $f8, 0x7($8)                           \r\n"
+        "gssdrc1 $f8, 0x0($8)                           \r\n"
+        "gsldlc1 $f16, 0x7($14)                         \r\n"
+        "gsldrc1 $f16, 0x0($14)                         \r\n"
+        "daddu $8, $10, $12                             \r\n"
+        "punpckhwd $f0, $f6, $f16                       \r\n"
+        "punpcklwd $f6, $f6, $f16                       \r\n"
+        "punpckhwd $f10, $f12, $f2                      \r\n"
+        "punpcklwd $f12, $f12, $f2                      \r\n"
+        "gssdlc1 $f6, 0x7($8)                           \r\n"
+        "gssdrc1 $f6, 0x0($8)                           \r\n"
+        "daddu $8, $9, $11                              \r\n"
+        "gssdlc1 $f0, 0x7($9)                           \r\n"
+        "gssdrc1 $f0, 0x0($9)                           \r\n"
+        "daddu $14, $9, $13                             \r\n"
+        "gssdlc1 $f12, 0x7($8)                          \r\n"
+        "gssdrc1 $f12, 0x0($8)                          \r\n"
+        "daddu $8, $13, $13                             \r\n"
+        "gssdlc1 $f10, 0x7($14)                         \r\n"
+        "gssdrc1 $f10, 0x0($14)                         \r\n"
+        "dsubu $10, $10, $8                             \r\n"
+        "dsubu $9, $9, $8                               \r\n"
+        "ldc1 $f0, 0x0+%[ptmp]                          \r\n"
+        "ldc1 $f2, 0x10+%[ptmp]                         \r\n"
+        "ldc1 $f4, 0x20+%[ptmp]                         \r\n"
+        "ldc1 $f6, 0x30+%[ptmp]                         \r\n"
+        "ldc1 $f8, 0x40+%[ptmp]                         \r\n"
+        "ldc1 $f10, 0x50+%[ptmp]                        \r\n"
+        "ldc1 $f12, 0x60+%[ptmp]                        \r\n"
+        "punpckhbh $f14, $f0, $f2                       \r\n"
+        "punpcklbh $f0, $f0, $f2                        \r\n"
+        "punpckhbh $f2, $f4, $f6                        \r\n"
+        "punpcklbh $f4, $f4, $f6                        \r\n"
+        "punpckhbh $f6, $f8, $f10                       \r\n"
+        "punpcklbh $f8, $f8, $f10                       \r\n"
+        "ldc1 $f16, 0x70+%[ptmp]                        \r\n"
+        "punpckhbh $f10, $f12, $f16                     \r\n"
+        "punpcklbh $f12, $f12, $f16                     \r\n"
+        "gssdlc1 $f6, 0x7($10)                          \r\n"
+        "gssdrc1 $f6, 0x0($10)                          \r\n"
+        "daddu $8, $10, $12                             \r\n"
+        "punpckhhw $f6, $f0, $f4                        \r\n"
+        "punpcklhw $f0, $f0, $f4                        \r\n"
+        "punpckhhw $f4, $f8, $f12                       \r\n"
+        "punpcklhw $f8, $f8, $f12                       \r\n"
+        "punpckhhw $f12, $f14, $f2                      \r\n"
+        "punpcklhw $f14, $f14, $f2                      \r\n"
+        "gssdlc1 $f4, 0x7($8)                           \r\n"
+        "gssdrc1 $f4, 0x0($8)                           \r\n"
+        "gsldlc1 $f4, 0x7($10)                          \r\n"
+        "gsldrc1 $f4, 0x0($10)                          \r\n"
+        "punpckhhw $f2, $f4, $f10                       \r\n"
+        "punpcklhw $f4, $f4, $f10                       \r\n"
+        "punpckhwd $f10, $f0, $f8                       \r\n"
+        "punpcklwd $f0, $f0, $f8                        \r\n"
+        "punpckhwd $f8, $f14, $f4                       \r\n"
+        "punpcklwd $f14, $f14, $f4                      \r\n"
+        "daddu $8, $10, %[stride]                       \r\n"
+        "gssdlc1 $f0, 0x7($10)                          \r\n"
+        "gssdrc1 $f0, 0x0($10)                          \r\n"
+        "daddu $14, $9, %[stride]                       \r\n"
+        "gssdlc1 $f10, 0x7($8)                          \r\n"
+        "gssdrc1 $f10, 0x0($8)                          \r\n"
+        "daddu $8, $9, $12                              \r\n"
+        "gssdlc1 $f14, 0x7($14)                         \r\n"
+        "gssdrc1 $f14, 0x0($14)                         \r\n"
+        "daddu $14, $10, $12                            \r\n"
+        "gssdlc1 $f8, 0x7($8)                           \r\n"
+        "gssdrc1 $f8, 0x0($8)                           \r\n"
+        "gsldlc1 $f16, 0x7($14)                         \r\n"
+        "gsldrc1 $f16, 0x0($14)                         \r\n"
+        "daddu $8, $10, $12                             \r\n"
+        "punpckhwd $f0, $f6, $f16                       \r\n"
+        "punpcklwd $f6, $f6, $f16                       \r\n"
+        "punpckhwd $f10, $f12, $f2                      \r\n"
+        "punpcklwd $f12, $f12, $f2                      \r\n"
+        "gssdlc1 $f6, 0x7($8)                           \r\n"
+        "gssdrc1 $f6, 0x0($8)                           \r\n"
+        "daddu $8, $9, $11                              \r\n"
+        "gssdlc1 $f0, 0x7($9)                           \r\n"
+        "gssdrc1 $f0, 0x0($9)                           \r\n"
+        "daddu $14, $9, $13                             \r\n"
+        "gssdlc1 $f12, 0x7($8)                          \r\n"
+        "gssdrc1 $f12, 0x0($8)                          \r\n"
+        "gssdlc1 $f10, 0x7($14)                         \r\n"
+        "gssdrc1 $f10, 0x0($14)                         \r\n"
+        ::[pix]"r"(pix),[stride]"r"((uint64_t)stride),[ptmp]"m"(ptmp[0]),
+          [pdat]"m"(pdat[0])
+        : "$8","$9","$10","$11","$12","$13","$14","$f0","$f2","$f4","$f6",
+          "$f8","$f10","$f12","$f14","$f16"
+    );
+}



More information about the ffmpeg-cvslog mailing list