;/****************************************************************************** ; * * ; * This file is part of XviD, a free MPEG-4 video encoder/decoder * ; * * ; * This program is free software; you can redistribute it and/or modify it * ; * under the terms of the GNU General Public License as published by * ; * the Free Software Foundation; either version 2 of the License, or * ; * (at your option) any later version. * ; * * ; * This program is distributed in the hope that it will be useful, but * ; * WITHOUT ANY WARRANTY; without even the implied warranty of * ; * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * ; * GNU General Public License for more details. * ; * * ; * You should have received a copy of the GNU General Public License * ; * along with this program; if not, write to the Free Software * ; * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA * ; * * ; ******************************************************************************/ ; ;/****************************************************************************** ; * * ; * fdct_mmx.asm, MMX optimized forward DCT * ; * * ; * Initial, but incomplete version provided by Intel at AppNote AP-922 * ; * http://developer.intel.com/vtune/cbts/strmsimd/922down.htm * ; * Copyright (C) 1999 Intel Corporation, * ; * * ; * completed and corrected in fdctmm32.c/fdctmm32.doc, * ; * http://members.tripod.com/~liaor * ; * Copyright (C) 2000 - Royce Shih-Wea Liao , * ; * * ; * ported to NASM and some minor changes * ; * Copyright (C) 2001 - Michael Militzer * ; * * ; * For more information visit the XviD homepage: http://www.xvid.org * ; * * ; ******************************************************************************/ ; ;/****************************************************************************** ; * * ; * Revision history: * ; * * ; * 04.11.2001 loop unrolled (Isibaar) * ; * 02.11.2001 initial version (Isibaar) * ; * * ; ******************************************************************************/ BITS 32 %macro cglobal 1 %ifdef PREFIX global _%1 %define %1 _%1 %else global %1 %endif %endmacro %define INP eax %define TABLE ebx %define TABLEF ebx %define OUT ecx %define round_frw_row edx %define INP_1 eax + 16 %define INP_2 eax + 32 %define INP_3 eax + 48 %define INP_4 eax + 64 %define INP_5 eax + 80 %define INP_6 eax + 96 %define INP_7 eax + 112 %define OUT_1 ecx + 16 %define OUT_2 ecx + 32 %define OUT_3 ecx + 48 %define OUT_4 ecx + 64 %define OUT_5 ecx + 80 %define OUT_6 ecx + 96 %define OUT_7 ecx + 112 %define OUT_8 ecx + 128 %define TABLE_1 ebx + 64 %define TABLE_2 ebx + 128 %define TABLE_3 ebx + 192 %define TABLE_4 ebx + 256 %define TABLE_5 ebx + 320 %define TABLE_6 ebx + 384 %define TABLE_7 ebx + 448 %define x0 INP + 0*16 %define x1 INP + 1*16 %define x2 INP + 2*16 %define x3 INP + 3*16 %define x4 INP + 4*16 %define x5 INP + 5*16 %define x6 INP + 6*16 %define x7 INP + 7*16 %define y0 OUT + 0*16 %define y1 OUT + 1*16 %define y2 OUT + 2*16 %define y3 OUT + 3*16 %define y4 OUT + 4*16 %define y5 OUT + 5*16 %define y6 OUT + 6*16 %define y7 OUT + 7*16 %define tg_1_16 (TABLEF + 0) %define tg_2_16 (TABLEF + 8) %define tg_3_16 (TABLEF + 16) %define cos_4_16 (TABLEF + 24) %define ocos_4_16 (TABLEF + 32) ;---added for sse2 %define tg_1_16_2 (TABLEF + 0) %define tg_2_16_2 (TABLEF + 16) %define tg_3_16_2 (TABLEF + 32) %define cos_4_16_2 (TABLEF + 48) %define ocos_4_16_2 (TABLEF + 64) ;--- %ifdef FORMAT_COFF SECTION .data data %else SECTION .data data align=64 %endif ALIGN 16 BITS_FRW_ACC equ 3 ; 2 or 3 for accuracy SHIFT_FRW_COL equ BITS_FRW_ACC SHIFT_FRW_ROW equ (BITS_FRW_ACC + 17) RND_FRW_ROW equ (1 << (SHIFT_FRW_ROW-1)) SHIFT_FRW_ROW_CLIP2 equ (4) SHIFT_FRW_ROW_CLIP1 equ (SHIFT_FRW_ROW - SHIFT_FRW_ROW_CLIP2) PASS1_BITS equ (2) CONST_BITS equ (13) one_corr dw 1, 1, 1, 1 ;---added for sse2 align 16 one_corr_2 dw 1, 1, 1, 1, dw 1, 1, 1, 1 ;--- r_frw_row dd RND_FRW_ROW, RND_FRW_ROW tg_all_16 dw 13036, 13036, 13036, 13036, ; tg * (2<<16) + 0.5 dw 27146, 27146, 27146, 27146, ; tg * (2<<16) + 0.5 dw -21746, -21746, -21746, -21746, ; tg * (2<<16) + 0.5 dw -19195, -19195, -19195, -19195, ; cos * (2<<16) + 0.5 dw 23170, 23170, 23170, 23170 ; cos * (2<<15) + 0.5 ;---added for sse2 align 16 tg_all_16_2 dw 13036, 13036, 13036, 13036, ; tg * (2<<16) + 0.5 dw 13036, 13036, 13036, 13036, dw 27146, 27146, 27146, 27146, ; tg * (2<<16) + 0.5 dw 27146, 27146, 27146, 27146, dw -21746, -21746, -21746, -21746, ; tg * (2<<16) + 0.5 dw -21746, -21746, -21746, -21746, dw -19195, -19195, -19195, -19195, ; cos * (2<<16) + 0.5 dw -19195, -19195, -19195, -19195, dw 23170, 23170, 23170, 23170 ; cos * (2<<15) + 0.5 dw 23170, 23170, 23170, 23170 ;--- tab_frw_01234567 ; row0 dw 16384, 16384, 21407, -8867, ; w09 w01 w08 w00 dw 16384, 16384, 8867, -21407, ; w13 w05 w12 w04 dw 16384, -16384, 8867, 21407, ; w11 w03 w10 w02 dw -16384, 16384, -21407, -8867, ; w15 w07 w14 w06 dw 22725, 12873, 19266, -22725, ; w22 w20 w18 w16 dw 19266, 4520, -4520, -12873, ; w23 w21 w19 w17 dw 12873, 4520, 4520, 19266, ; w30 w28 w26 w24 dw -22725, 19266, -12873, -22725, ; w31 w29 w27 w25 ; row1 dw 22725, 22725, 29692, -12299, ; w09 w01 w08 w00 dw 22725, 22725, 12299, -29692, ; w13 w05 w12 w04 dw 22725, -22725, 12299, 29692, ; w11 w03 w10 w02 dw -22725, 22725, -29692, -12299, ; w15 w07 w14 w06 dw 31521, 17855, 26722, -31521, ; w22 w20 w18 w16 dw 26722, 6270, -6270, -17855, ; w23 w21 w19 w17 dw 17855, 6270, 6270, 26722, ; w30 w28 w26 w24 dw -31521, 26722, -17855, -31521, ; w31 w29 w27 w25 ; row2 dw 21407, 21407, 27969, -11585, ; w09 w01 w08 w00 dw 21407, 21407, 11585, -27969, ; w13 w05 w12 w04 dw 21407, -21407, 11585, 27969, ; w11 w03 w10 w02 dw -21407, 21407, -27969, -11585, ; w15 w07 w14 w06 dw 29692, 16819, 25172, -29692, ; w22 w20 w18 w16 dw 25172, 5906, -5906, -16819, ; w23 w21 w19 w17 dw 16819, 5906, 5906, 25172, ; w30 w28 w26 w24 dw -29692, 25172, -16819, -29692, ; w31 w29 w27 w25 ; row3 dw 19266, 19266, 25172, -10426, ; w09 w01 w08 w00 dw 19266, 19266, 10426, -25172, ; w13 w05 w12 w04 dw 19266, -19266, 10426, 25172, ; w11 w03 w10 w02 dw -19266, 19266, -25172, -10426, ; w15 w07 w14 w06 dw 26722, 15137, 22654, -26722, ; w22 w20 w18 w16 dw 22654, 5315, -5315, -15137, ; w23 w21 w19 w17 dw 15137, 5315, 5315, 22654, ; w30 w28 w26 w24 dw -26722, 22654, -15137, -26722, ; w31 w29 w27 w25 ; row4 dw 16384, 16384, 21407, -8867, ; w09 w01 w08 w00 dw 16384, 16384, 8867, -21407, ; w13 w05 w12 w04 dw 16384, -16384, 8867, 21407, ; w11 w03 w10 w02 dw -16384, 16384, -21407, -8867, ; w15 w07 w14 w06 dw 22725, 12873, 19266, -22725, ; w22 w20 w18 w16 dw 19266, 4520, -4520, -12873, ; w23 w21 w19 w17 dw 12873, 4520, 4520, 19266, ; w30 w28 w26 w24 dw -22725, 19266, -12873, -22725, ; w31 w29 w27 w25 ; row5 dw 19266, 19266, 25172, -10426, ; w09 w01 w08 w00 dw 19266, 19266, 10426, -25172, ; w13 w05 w12 w04 dw 19266, -19266, 10426, 25172, ; w11 w03 w10 w02 dw -19266, 19266, -25172, -10426, ; w15 w07 w14 w06 dw 26722, 15137, 22654, -26722, ; w22 w20 w18 w16 dw 22654, 5315, -5315, -15137, ; w23 w21 w19 w17 dw 15137, 5315, 5315, 22654, ; w30 w28 w26 w24 dw -26722, 22654, -15137, -26722, ; w31 w29 w27 w25 ; row6 dw 21407, 21407, 27969, -11585, ; w09 w01 w08 w00 dw 21407, 21407, 11585, -27969, ; w13 w05 w12 w04 dw 21407, -21407, 11585, 27969, ; w11 w03 w10 w02 dw -21407, 21407, -27969, -11585, ; w15 w07 w14 w06 dw 29692, 16819, 25172, -29692, ; w22 w20 w18 w16 dw 25172, 5906, -5906, -16819, ; w23 w21 w19 w17 dw 16819, 5906, 5906, 25172, ; w30 w28 w26 w24 dw -29692, 25172, -16819, -29692, ; w31 w29 w27 w25 ; row7 dw 22725, 22725, 29692, -12299, ; w09 w01 w08 w00 dw 22725, 22725, 12299, -29692, ; w13 w05 w12 w04 dw 22725, -22725, 12299, 29692, ; w11 w03 w10 w02 dw -22725, 22725, -29692, -12299, ; w15 w07 w14 w06 dw 31521, 17855, 26722, -31521, ; w22 w20 w18 w16 dw 26722, 6270, -6270, -17855, ; w23 w21 w19 w17 dw 17855, 6270, 6270, 26722, ; w30 w28 w26 w24 dw -31521, 26722, -17855, -31521 ; w31 w29 w27 w25 align 128 FIX_1 dw 10703, 4433, 10703, 4433 dw 10703, 4433, 10703, 4433 FIX_2 dw 4433, -10704, 4433, -10704 dw 4433, -10704, 4433, -10704 rounder_5 dw 2+16,2+16,2+16,2+16,2+16,2+16,2+16,2+16 rounder_11 dd 1024,1024,1024,1024 rounder_18 dd 16384+131072,16384+131072,16384+131072,16384+131072 align 128 FIX_3 dw 6437, 2260, 6437, 2260 dw 6437, 2260, 6437, 2260 dw 11363, 9633, 11363, 9633 dw 11363, 9633, 11363, 9633 FIX_4 dw -11362, -6436, -11362, -6436 dw -11362, -6436, -11362, -6436 dw 9633, -2259, 9633, -2259 dw 9633, -2259, 9633, -2259 FIX_5 dw 2261, 9633, 2261, 9633 dw 2261, 9633, 2261, 9633 dw 6437, -11362, 6437, -11362 dw 6437, -11362, 6437, -11362 FIX_6 dw 9633, -11363, 9633, -11363 dw 9633, -11363, 9633, -11363 dw 2260, -6436, 2260, -6436 dw 2260, -6436, 2260, -6436 align 128 buffer dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 dw 0,0,0,0,0,0,0,0 SECTION .text ALIGN 16 cglobal fdct_mmx ;;void fdct_mmx(short *block); fdct_mmx: push ebx mov INP, dword [esp + 8] ; block mov TABLEF, tg_all_16 mov OUT, INP movq mm0, [x1] ; 0 ; x1 movq mm1, [x6] ; 1 ; x6 movq mm2, mm0 ; 2 ; x1 movq mm3, [x2] ; 3 ; x2 paddsw mm0, mm1 ; t1 = x[1] + x[6] movq mm4, [x5] ; 4 ; x5 psllw mm0, SHIFT_FRW_COL ; t1 movq mm5, [x0] ; 5 ; x0 paddsw mm4, mm3 ; t2 = x[2] + x[5] paddsw mm5, [x7] ; t0 = x[0] + x[7] psllw mm4, SHIFT_FRW_COL ; t2 movq mm6, mm0 ; 6 ; t1 psubsw mm2, mm1 ; 1 ; t6 = x[1] - x[6] movq mm1, [tg_2_16] ; 1 ; tg_2_16 psubsw mm0, mm4 ; tm12 = t1 - t2 movq mm7, [x3] ; 7 ; x3 pmulhw mm1, mm0 ; tm12*tg_2_16 paddsw mm7, [x4] ; t3 = x[3] + x[4] psllw mm5, SHIFT_FRW_COL ; t0 paddsw mm6, mm4 ; 4 ; tp12 = t1 + t2 psllw mm7, SHIFT_FRW_COL ; t3 movq mm4, mm5 ; 4 ; t0 psubsw mm5, mm7 ; tm03 = t0 - t3 paddsw mm1, mm5 ; y2 = tm03 + tm12*tg_2_16 paddsw mm4, mm7 ; 7 ; tp03 = t0 + t3 por mm1, qword [one_corr] ; correction y2 +0.5 psllw mm2, SHIFT_FRW_COL+1 ; t6 pmulhw mm5, [tg_2_16] ; tm03*tg_2_16 movq mm7, mm4 ; 7 ; tp03 psubsw mm3, [x5] ; t5 = x[2] - x[5] psubsw mm4, mm6 ; y4 = tp03 - tp12 movq [y2], mm1 ; 1 ; save y2 paddsw mm7, mm6 ; 6 ; y0 = tp03 + tp12 movq mm1, [x3] ; 1 ; x3 psllw mm3, SHIFT_FRW_COL+1 ; t5 psubsw mm1, [x4] ; t4 = x[3] - x[4] movq mm6, mm2 ; 6 ; t6 movq [y4], mm4 ; 4 ; save y4 paddsw mm2, mm3 ; t6 + t5 pmulhw mm2, [ocos_4_16] ; tp65 = (t6 + t5)*cos_4_16 psubsw mm6, mm3 ; 3 ; t6 - t5 pmulhw mm6, [ocos_4_16] ; tm65 = (t6 - t5)*cos_4_16 psubsw mm5, mm0 ; 0 ; y6 = tm03*tg_2_16 - tm12 por mm5, qword [one_corr] ; correction y6 +0.5 psllw mm1, SHIFT_FRW_COL ; t4 por mm2, qword [one_corr] ; correction tp65 +0.5 movq mm4, mm1 ; 4 ; t4 movq mm3, [x0] ; 3 ; x0 paddsw mm1, mm6 ; tp465 = t4 + tm65 psubsw mm3, [x7] ; t7 = x[0] - x[7] psubsw mm4, mm6 ; 6 ; tm465 = t4 - tm65 movq mm0, [tg_1_16] ; 0 ; tg_1_16 psllw mm3, SHIFT_FRW_COL ; t7 movq mm6, [tg_3_16] ; 6 ; tg_3_16 pmulhw mm0, mm1 ; tp465*tg_1_16 movq [y0], mm7 ; 7 ; save y0 pmulhw mm6, mm4 ; tm465*tg_3_16 movq [y6], mm5 ; 5 ; save y6 movq mm7, mm3 ; 7 ; t7 movq mm5, [tg_3_16] ; 5 ; tg_3_16 psubsw mm7, mm2 ; tm765 = t7 - tp65 paddsw mm3, mm2 ; 2 ; tp765 = t7 + tp65 pmulhw mm5, mm7 ; tm765*tg_3_16 paddsw mm0, mm3 ; y1 = tp765 + tp465*tg_1_16 paddsw mm6, mm4 ; tm465*tg_3_16 pmulhw mm3, [tg_1_16] ; tp765*tg_1_16 por mm0, qword [one_corr] ; correction y1 +0.5 paddsw mm5, mm7 ; tm765*tg_3_16 psubsw mm7, mm6 ; 6 ; y3 = tm765 - tm465*tg_3_16 add INP, 0x08 movq [y1], mm0 ; 0 ; save y1 paddsw mm5, mm4 ; 4 ; y5 = tm765*tg_3_16 + tm465 movq [y3], mm7 ; 7 ; save y3 psubsw mm3, mm1 ; 1 ; y7 = tp765*tg_1_16 - tp465 movq [y5], mm5 ; 5 ; save y5 movq mm0, [x1] ; 0 ; x1 movq [y7], mm3 ; 3 ; save y7 (columns 0-4) movq mm1, [x6] ; 1 ; x6 movq mm2, mm0 ; 2 ; x1 movq mm3, [x2] ; 3 ; x2 paddsw mm0, mm1 ; t1 = x[1] + x[6] movq mm4, [x5] ; 4 ; x5 psllw mm0, SHIFT_FRW_COL ; t1 movq mm5, [x0] ; 5 ; x0 paddsw mm4, mm3 ; t2 = x[2] + x[5] paddsw mm5, [x7] ; t0 = x[0] + x[7] psllw mm4, SHIFT_FRW_COL ; t2 movq mm6, mm0 ; 6 ; t1 psubsw mm2, mm1 ; 1 ; t6 = x[1] - x[6] movq mm1, [tg_2_16] ; 1 ; tg_2_16 psubsw mm0, mm4 ; tm12 = t1 - t2 movq mm7, [x3] ; 7 ; x3 pmulhw mm1, mm0 ; tm12*tg_2_16 paddsw mm7, [x4] ; t3 = x[3] + x[4] psllw mm5, SHIFT_FRW_COL ; t0 paddsw mm6, mm4 ; 4 ; tp12 = t1 + t2 psllw mm7, SHIFT_FRW_COL ; t3 movq mm4, mm5 ; 4 ; t0 psubsw mm5, mm7 ; tm03 = t0 - t3 paddsw mm1, mm5 ; y2 = tm03 + tm12*tg_2_16 paddsw mm4, mm7 ; 7 ; tp03 = t0 + t3 por mm1, qword [one_corr] ; correction y2 +0.5 psllw mm2, SHIFT_FRW_COL+1 ; t6 pmulhw mm5, [tg_2_16] ; tm03*tg_2_16 movq mm7, mm4 ; 7 ; tp03 psubsw mm3, [x5] ; t5 = x[2] - x[5] psubsw mm4, mm6 ; y4 = tp03 - tp12 movq [y2+8], mm1 ; 1 ; save y2 paddsw mm7, mm6 ; 6 ; y0 = tp03 + tp12 movq mm1, [x3] ; 1 ; x3 psllw mm3, SHIFT_FRW_COL+1 ; t5 psubsw mm1, [x4] ; t4 = x[3] - x[4] movq mm6, mm2 ; 6 ; t6 movq [y4+8], mm4 ; 4 ; save y4 paddsw mm2, mm3 ; t6 + t5 pmulhw mm2, [ocos_4_16] ; tp65 = (t6 + t5)*cos_4_16 psubsw mm6, mm3 ; 3 ; t6 - t5 pmulhw mm6, [ocos_4_16] ; tm65 = (t6 - t5)*cos_4_16 psubsw mm5, mm0 ; 0 ; y6 = tm03*tg_2_16 - tm12 por mm5, qword [one_corr] ; correction y6 +0.5 psllw mm1, SHIFT_FRW_COL ; t4 por mm2, qword [one_corr] ; correction tp65 +0.5 movq mm4, mm1 ; 4 ; t4 movq mm3, [x0] ; 3 ; x0 paddsw mm1, mm6 ; tp465 = t4 + tm65 psubsw mm3, [x7] ; t7 = x[0] - x[7] psubsw mm4, mm6 ; 6 ; tm465 = t4 - tm65 movq mm0, [tg_1_16] ; 0 ; tg_1_16 psllw mm3, SHIFT_FRW_COL ; t7 movq mm6, [tg_3_16] ; 6 ; tg_3_16 pmulhw mm0, mm1 ; tp465*tg_1_16 movq [y0+8], mm7 ; 7 ; save y0 pmulhw mm6, mm4 ; tm465*tg_3_16 movq [y6+8], mm5 ; 5 ; save y6 movq mm7, mm3 ; 7 ; t7 movq mm5, [tg_3_16] ; 5 ; tg_3_16 psubsw mm7, mm2 ; tm765 = t7 - tp65 paddsw mm3, mm2 ; 2 ; tp765 = t7 + tp65 pmulhw mm5, mm7 ; tm765*tg_3_16 paddsw mm0, mm3 ; y1 = tp765 + tp465*tg_1_16 paddsw mm6, mm4 ; tm465*tg_3_16 pmulhw mm3, [tg_1_16] ; tp765*tg_1_16 por mm0, qword [one_corr] ; correction y1 +0.5 paddsw mm5, mm7 ; tm765*tg_3_16 psubsw mm7, mm6 ; 6 ; y3 = tm765 - tm465*tg_3_16 movq [y1+8], mm0 ; 0 ; save y1 paddsw mm5, mm4 ; 4 ; y5 = tm765*tg_3_16 + tm465 movq [y3+8], mm7 ; 7 ; save y3 psubsw mm3, mm1 ; 1 ; y7 = tp765*tg_1_16 - tp465 movq [y5+8], mm5 ; 5 ; save y5 movq [y7+8], mm3 ; 3 ; save y7 ;--------------------- columns mov INP, [esp + 8] ; row 0 mov TABLEF, tab_frw_01234567 ; row 0 mov OUT, INP mov round_frw_row, r_frw_row movd mm5, [INP+12] ; mm5 = 7 6 punpcklwd mm5, [INP+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_1-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_1-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_1+12] ; mm5 = 7 6 punpcklwd mm5, [INP_1+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_1] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_1] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_1+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_1+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_1+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_1+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_1+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_1+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_1+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_2-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_2-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_2+12] ; mm5 = 7 6 punpcklwd mm5, [INP_2+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_2] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_2] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_2+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_2+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_2+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_2+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_2+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_2+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_2+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_3-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_3-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_3+12] ; mm5 = 7 6 punpcklwd mm5, [INP_3+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_3] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_3] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_3+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_3+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_3+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_3+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_3+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_3+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_3+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_4-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_4-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_4+12] ; mm5 = 7 6 punpcklwd mm5, [INP_4+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_4] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_4] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_4+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_4+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_4+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_4+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_4+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_4+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_4+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_5-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_5-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_5+12] ; mm5 = 7 6 punpcklwd mm5, [INP_5+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_5] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_5] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_5+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_5+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_5+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_5+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_5+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_5+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_5+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_6-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_6-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_6+12] ; mm5 = 7 6 punpcklwd mm5, [INP_6+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_6] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_6] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_6+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_6+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_6+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_6+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_6+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_6+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_6+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_7-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_7-8], mm6 ; 7 ; save y7 y6 y5 y4 movd mm5, [INP_7+12] ; mm5 = 7 6 punpcklwd mm5, [INP_7+8] movq mm2, mm5 ; mm2 = 5 7 4 6 psrlq mm5, 32 ; mm5 = _ _ 5 7 movq mm0, [INP_7] ; mm0 = 3 2 1 0 punpcklwd mm5, mm2 ; mm5 = 4 5 6 7 movq mm1, mm0 ; mm1 = 3 2 1 0 paddsw mm0, mm5 ; mm0 = [3+4, 2+5, 1+6, 0+7] (xt3, xt2, xt1, xt0) psubsw mm1, mm5 ; mm1 = [3-4, 2-5, 1-6, 0-7] (xt7, xt6, xt5, xt4) movq mm2, mm0 ; mm2 = [ xt3 xt2 xt1 xt0 ] punpcklwd mm0, mm1 ; mm0 = [ xt5 xt1 xt4 xt0 ] punpckhwd mm2, mm1 ; mm2 = [ xt7 xt3 xt6 xt2 ] movq mm1, mm2 ; mm1 movq mm2, mm0 ; 2 ; x3 x2 x1 x0 movq mm3, [TABLE_7] ; 3 ; w06 w04 w02 w00 punpcklwd mm0, mm1 ; x5 x1 x4 x0 movq mm5, mm0 ; 5 ; x5 x1 x4 x0 punpckldq mm0, mm0 ; x4 x0 x4 x0 [ xt2 xt0 xt2 xt0 ] movq mm4, [TABLE_7+8] ; 4 ; w07 w05 w03 w01 punpckhwd mm2, mm1 ; 1 ; x7 x3 x6 x2 pmaddwd mm3, mm0 ; x4*w06+x0*w04 x4*w02+x0*w00 movq mm6, mm2 ; 6 ; x7 x3 x6 x2 movq mm1, [TABLE_7+32] ; 1 ; w22 w20 w18 w16 punpckldq mm2, mm2 ; x6 x2 x6 x2 [ xt3 xt1 xt3 xt1 ] pmaddwd mm4, mm2 ; x6*w07+x2*w05 x6*w03+x2*w01 punpckhdq mm5, mm5 ; x5 x1 x5 x1 [ xt6 xt4 xt6 xt4 ] pmaddwd mm0, [TABLE_7+16] ; x4*w14+x0*w12 x4*w10+x0*w08 punpckhdq mm6, mm6 ; x7 x3 x7 x3 [ xt7 xt5 xt7 xt5 ] movq mm7, [TABLE_7+40] ; 7 ; w23 w21 w19 w17 pmaddwd mm1, mm5 ; x5*w22+x1*w20 x5*w18+x1*w16 paddd mm3, [round_frw_row] ; +rounder (y2,y0) pmaddwd mm7, mm6 ; x7*w23+x3*w21 x7*w19+x3*w17 pmaddwd mm2, [TABLE_7+24] ; x6*w15+x2*w13 x6*w11+x2*w09 paddd mm3, mm4 ; 4 ; a1=sum(even1) a0=sum(even0) pmaddwd mm5, [TABLE_7+48] ; x5*w30+x1*w28 x5*w26+x1*w24 pmaddwd mm6, [TABLE_7+56] ; x7*w31+x3*w29 x7*w27+x3*w25 paddd mm1, mm7 ; 7 ; b1=sum(odd1) b0=sum(odd0) paddd mm0, [round_frw_row] ; +rounder (y6,y4) psrad mm3, SHIFT_FRW_ROW_CLIP1 ; (y2, y0) paddd mm1, [round_frw_row] ; +rounder (y3,y1) paddd mm0, mm2 ; 2 ; a3=sum(even3) a2=sum(even2) paddd mm5, [round_frw_row] ; +rounder (y7,y5) psrad mm1, SHIFT_FRW_ROW_CLIP1 ; y1=a1+b1 y0=a0+b0 paddd mm5, mm6 ; 6 ; b3=sum(odd3) b2=sum(odd2) psrad mm0, SHIFT_FRW_ROW_CLIP1 ; y3=a3+b3 y2=a2+b2 psrad mm5, SHIFT_FRW_ROW_CLIP1 ; y4=a3-b3 y5=a2-b2 packssdw mm3, mm0 ; 0 ; y6 y4 y2 y0, saturate {-32768,+32767} packssdw mm1, mm5 ; 3 ; y7 y5 y3 y1, saturate {-32768,+32767} movq mm6, mm3 ; mm0 = y6 y4 y2 y0 punpcklwd mm3, mm1 ; y3 y2 y1 y0 punpckhwd mm6, mm1 ; y7 y6 y5 y4 psraw mm3, SHIFT_FRW_ROW_CLIP2 ; descale [y3 y2 y1 y0] to {-2048,+2047} psraw mm6, SHIFT_FRW_ROW_CLIP2 ; descale [y7 y6 y5 y4] to {-2048,+2047} movq [OUT_8-16], mm3 ; 1 ; save y3 y2 y1 y0 movq [OUT_8-8], mm6 ; 7 ; save y7 y6 y5 y4 pop ebx emms ret ;******************************************************************* ; This SSE2 code of the FDCT algoruthm coded by ; Dmitry Rozhdestvensky and checked by Vladimir G. Ivanov ;******************************************************************* %macro transpose8x8sse2 3 ; source,dest,{0=first part, 1=second part} ;I found this smart transposing algo at www.x86.org movdqa xmm0,[%1+0*16] ; 07 06 05 04 03 02 01 00 movdqa xmm6,[%1+2*16] ; 27 26 25 24 23 22 21 20 movdqa xmm4,[%1+4*16] ; 47 46 45 44 43 42 41 40 movdqa xmm7,[%1+6*16] ; 67 66 65 64 63 62 61 60 %if %3=0 punpcklwd xmm0,[%1+1*16] ; 13 03 12 02 11 01 10 00 movdqa xmm2,xmm0 punpcklwd xmm6,[%1+3*16] ; 33 23 32 22 31 21 30 20 punpcklwd xmm4,[%1+5*16] ; 53 43 52 42 51 41 50 40 movdqa xmm5,xmm4 punpcklwd xmm7,[%1+7*16] ; 73 63 72 62 71 61 70 60 %else punpckhwd xmm0,[%1+1*16] ; movdqa xmm2,xmm0 punpckhwd xmm6,[%1+3*16] ; punpckhwd xmm4,[%1+5*16] ; movdqa xmm5,xmm4 punpckhwd xmm7,[%1+7*16] ; %endif punpckldq xmm0,xmm6 ; 31 21 11 01 30 20 10 00 movdqa xmm1,xmm0 punpckldq xmm4,xmm7 ; 71 61 51 41 70 60 50 40 punpckhdq xmm2,xmm6 ; 33 23 13 03 32 22 12 02 movdqa xmm3,xmm2 punpckhdq xmm5,xmm7 ; 73 63 53 43 72 62 52 42 punpcklqdq xmm0,xmm4 ; 70 60 50 40 30 20 10 00 punpcklqdq xmm2,xmm5 ; 72 62 52 42 32 22 21 02 punpckhqdq xmm1,xmm4 ; 71 61 51 41 31 21 11 01 punpckhqdq xmm3,xmm5 ; 73 63 53 43 33 23 13 03 movdqa [%2+(0+%3*4)*16],xmm0 movdqa [%2+(1+%3*4)*16],xmm1 movdqa [%2+(2+%3*4)*16],xmm2 movdqa [%2+(3+%3*4)*16],xmm3 %endmacro %macro makeoddpart 3 ; output, table, shift movdqa xmm1,[%2+0 ] ;45l movdqa xmm5,[%2+16] ;67l movdqa xmm3,xmm1 ;45h movdqa xmm7,xmm5 ;67h pmaddwd xmm1,xmm2 ;[%2+0 ] pmaddwd xmm5,xmm0 ;[%2+16] paddd xmm1,xmm5 %if %3=11 movdqa xmm5,[rounder_11] %else movdqa xmm5,[rounder_18] %endif pmaddwd xmm3,xmm6 ;[%2+0 ] pmaddwd xmm7,xmm4 ;[%2+16] paddd xmm3,xmm7 paddd xmm1,xmm5 ;rounder paddd xmm3,xmm5 ;rounder psrad xmm1,%3 psrad xmm3,%3 packssdw xmm1,xmm3 movdqa [%1],xmm1 %endmacro %macro makeoddpartlast 3 ; output, table, shift pmaddwd xmm2,[%2+0 ] pmaddwd xmm0,[%2+16] paddd xmm2,xmm0 pmaddwd xmm6,[%2+0 ] pmaddwd xmm4,[%2+16] paddd xmm6,xmm4 paddd xmm2,xmm5 ;rounder paddd xmm6,xmm5 ;rounder psrad xmm2,%3 psrad xmm6,%3 packssdw xmm2,xmm6 movdqa [%1],xmm2 %endmacro %macro FDCT_1D 4 ; INP,OUTP,{0=first pass, 1=second pass}, shift={11,18} ;movdqa xmm0,[%1+16*0] ;We do not load 0-3 values here for they ;movdqa xmm1,[%1+16*1] ;stayed from transposition ;movdqa xmm2,[%1+16*2] ;movdqa xmm3,[%1+16*3] %if %3<>0 movdqa xmm7,[rounder_5] %endif paddsw xmm0,[%1+16*7] ;tmp0 movdqa xmm4,xmm0 paddsw xmm1,[%1+16*6] ;tmp1 movdqa xmm5,xmm1 paddsw xmm2,[%1+16*5] ;tmp2 paddsw xmm3,[%1+16*4] ;tmp3 paddsw xmm0,xmm3 ;tmp10 %if %3<>0 ; In the second pass we must round and shift before ; the tmp10+tmp11 and tmp10-tmp11 calculation ; or the overflow will happen. paddsw xmm0,xmm7 ;[rounder_5] psraw xmm0,PASS1_BITS+3 %endif movdqa xmm6,xmm0 ;tmp10 paddsw xmm1,xmm2 ;tmp11 psubsw xmm4,xmm3 ;tmp13 psubsw xmm5,xmm2 ;tmp12 %if %3=0 paddsw xmm0,xmm1 psubsw xmm6,xmm1 psllw xmm0,PASS1_BITS psllw xmm6,PASS1_BITS %else paddsw xmm1,xmm7 ;[rounder_5] psraw xmm1,PASS1_BITS+3 paddsw xmm0,xmm1 psubsw xmm6,xmm1 %endif movdqa xmm1,xmm4 movdqa xmm2,xmm4 movdqa [%2+16*0],xmm0 movdqa [%2+16*4],xmm6 movdqa xmm7,[FIX_1] punpckhwd xmm1,xmm5 ; 12 13 12 13 12 13 12 13 high part movdqa xmm6,xmm1 ;high punpcklwd xmm2,xmm5 ; 12 13 12 13 12 13 12 13 low part movdqa xmm0,xmm2 ;low movdqa xmm4,[FIX_2] %if %4=11 movdqa xmm5,[rounder_11] %else movdqa xmm5,[rounder_18] %endif pmaddwd xmm2,xmm7 ;[FIX_1] pmaddwd xmm1,xmm7 ;[FIX_1] pmaddwd xmm0,xmm4 ;[FIX_2] pmaddwd xmm6,xmm4 ;[FIX_2] paddd xmm2,xmm5 ;rounder paddd xmm1,xmm5 ;rounder psrad xmm2,%4 psrad xmm1,%4 packssdw xmm2,xmm1 movdqa [%2+16*2],xmm2 paddd xmm0,xmm5 ;rounder paddd xmm6,xmm5 ;rounder psrad xmm0,%4 psrad xmm6,%4 packssdw xmm0,xmm6 movdqa [%2+16*6],xmm0 movdqa xmm0,[%1+16*0] movdqa xmm1,[%1+16*1] movdqa xmm2,[%1+16*2] movdqa xmm3,[%1+16*3] psubsw xmm0,[%1+16*7] ;tmp7 movdqa xmm4,xmm0 psubsw xmm1,[%1+16*6] ;tmp6 psubsw xmm2,[%1+16*5] ;tmp5 movdqa xmm6,xmm2 psubsw xmm3,[%1+16*4] ;tmp4 punpckhwd xmm4,xmm1 ; 6 7 6 7 6 7 6 7 high part punpcklwd xmm0,xmm1 ; 6 7 6 7 6 7 6 7 low part punpckhwd xmm6,xmm3 ; 4 5 4 5 4 5 4 5 high part punpcklwd xmm2,xmm3 ; 4 5 4 5 4 5 4 5 low part makeoddpart %2+16*1,FIX_3,%4 makeoddpart %2+16*3,FIX_4,%4 makeoddpart %2+16*5,FIX_5,%4 makeoddpartlast %2+16*7,FIX_6,%4 %endmacro cglobal fdct_sse2 ;;void f dct_sse2(short *block); fdct_sse2: push eax push ebx mov eax,[esp+4+2*4] mov ebx,buffer prefetchnta [FIX_1] prefetchnta [FIX_3] prefetchnta [FIX_5] transpose8x8sse2 eax,ebx,1 ; First we transpose last 4 lines transpose8x8sse2 eax,ebx,0 ; Then the firts 4 lines ;processing columns (became rows after transposition) FDCT_1D ebx,eax,0,CONST_BITS - PASS1_BITS transpose8x8sse2 eax,ebx,1 transpose8x8sse2 eax,ebx,0 ;now processing rows FDCT_1D ebx,eax,1,CONST_BITS + PASS1_BITS + 3 pop ebx pop eax ret