;/***************************************************************************** ; * ; * XVID MPEG-4 VIDEO CODEC ; * mmx yuv planar to yv12 conversion ; * ; * Copyright (C) 2001 - Michael Militzer ; * ; * This file is part of XviD, a free MPEG-4 video encoder/decoder ; * ; * XviD is free software; you can redistribute it and/or modify it ; * under the terms of the GNU General Public License as published by ; * the Free Software Foundation; either version 2 of the License, or ; * (at your option) any later version. ; * ; * This program is distributed in the hope that it will be useful, ; * but WITHOUT ANY WARRANTY; without even the implied warranty of ; * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ; * GNU General Public License for more details. ; * ; * You should have received a copy of the GNU General Public License ; * along with this program; if not, write to the Free Software ; * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ; * ; * Under section 8 of the GNU General Public License, the copyright ; * holders of XVID explicitly forbid distribution in the following ; * countries: ; * ; * - Japan ; * - United States of America ; * ; * Linking XviD statically or dynamically with other modules is making a ; * combined work based on XviD. Thus, the terms and conditions of the ; * GNU General Public License cover the whole combination. ; * ; * As a special exception, the copyright holders of XviD give you ; * permission to link XviD with independent modules that communicate with ; * XviD solely through the VFW1.1 and DShow interfaces, regardless of the ; * license terms of these independent modules, and to copy and distribute ; * the resulting combined work under terms of your choice, provided that ; * every copy of the combined work is accompanied by a complete copy of ; * the source code of XviD (the version of XviD used to produce the ; * combined work), being distributed under the terms of the GNU General ; * Public License plus this exception. An independent module is a module ; * which is not derived from or based on XviD. ; * ; * Note that people who make modified versions of XviD are not obligated ; * to grant this special exception for their modified versions; it is ; * their choice whether to do so. The GNU General Public License gives ; * permission to release a modified version without this exception; this ; * exception also makes it possible to release a modified version which ; * carries forward this exception. ; * ; * $Id: yuv_to_yv12_mmx.asm,v 1.7 2002-11-17 00:20:30 edgomez Exp $ ; * ; ****************************************************************************/ BITS 32 %macro cglobal 1 %ifdef PREFIX global _%1 %define %1 _%1 %else global %1 %endif %endmacro SECTION .text ALIGN 64 ;------------------------------------------------------------------------------ ; ; void yuv_to_yv12_xmm(uint8_t *y_out, ; uint8_t *u_out, ; uint8_t *v_out, ; uint8_t *src, ; int width, int height, int stride); ; ; This function probably also runs on PentiumII class cpu's ; ; Attention: This code assumes that width is a multiple of 16 ; ;------------------------------------------------------------------------------ cglobal yuv_to_yv12_xmm yuv_to_yv12_xmm: push ebx push esi push edi push ebp ; local vars allocation %define localsize 4 %define remainder esp sub esp, localsize ; function code mov eax, [esp + 40 + localsize] ; height -> eax mov ebx, [esp + 44 + localsize] ; stride -> ebx mov esi, [esp + 32 + localsize] ; src -> esi mov edi, [esp + 20 + localsize] ; y_out -> edi mov ecx, [esp + 36 + localsize] ; width -> ecx sub ebx, ecx ; stride - width -> ebx mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; remainder -> ebp shr ebp, 4 ; 16 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .y_inner_loop: prefetchnta [esi + 64] ; non temporal prefetch prefetchnta [esi + 96] movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movntq [edi], mm1 ; write to y_out movntq [edi + 8], mm2 movntq [edi + 16], mm3 movntq [edi + 24], mm4 movntq [edi + 32], mm5 movntq [edi + 40], mm6 movntq [edi + 48], mm7 movntq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .y_inner_loop dec ebp jz .y_outer_loop .y_remainder_loop: movq mm1, [esi] ; read from src movq mm2, [esi + 8] movntq [edi], mm1 ; write to y_out movntq [edi + 8], mm2 add esi, 16 add edi, 16 dec ebp jnz .y_remainder_loop .y_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz near .y_inner_loop mov eax, [esp + 40 + localsize] ; height -> eax mov ebx, [esp + 44 + localsize] ; stride -> ebx mov ecx, [esp + 36 + localsize] ; width -> ecx mov edi, [esp + 24 + localsize] ; u_out -> edi shr ecx, 1 ; width / 2 -> ecx shr ebx, 1 ; stride / 2 -> ebx shr eax, 1 ; height / 2 -> eax sub ebx, ecx ; stride / 2 - width / 2 -> ebx mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; remainder -> ebp shr ebp, 3 ; 8 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .u_inner_loop: prefetchnta [esi + 64] ; non temporal prefetch prefetchnta [esi + 96] movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movntq [edi], mm1 ; write to u_out movntq [edi + 8], mm2 movntq [edi + 16], mm3 movntq [edi + 24], mm4 movntq [edi + 32], mm5 movntq [edi + 40], mm6 movntq [edi + 48], mm7 movntq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .u_inner_loop dec ebp jz .u_outer_loop .u_remainder_loop: movq mm1, [esi] ; read from src movntq [edi], mm1 ; write to y_out add esi, 8 add edi, 8 dec ebp jnz .u_remainder_loop .u_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz .u_inner_loop mov eax, [esp + 40 + localsize] ; height -> eax mov ecx, [esp + 36 + localsize] ; width -> ecx mov edi, [esp + 28 + localsize] ; v_out -> edi shr ecx, 1 ; width / 2 -> ecx shr eax, 1 ; height / 2 -> eax mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; remainder -> ebp shr ebp, 3 ; 8 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .v_inner_loop: prefetchnta [esi + 64] ; non temporal prefetch prefetchnta [esi + 96] movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movntq [edi], mm1 ; write to u_out movntq [edi + 8], mm2 movntq [edi + 16], mm3 movntq [edi + 24], mm4 movntq [edi + 32], mm5 movntq [edi + 40], mm6 movntq [edi + 48], mm7 movntq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .v_inner_loop dec ebp jz .v_outer_loop .v_remainder_loop: movq mm1, [esi] ; read from src movntq [edi], mm1 ; write to y_out add esi, 8 add edi, 8 dec ebp jnz .v_remainder_loop .v_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz .v_inner_loop ; local vars deallocation add esp, localsize %undef localsize %undef remainder pop ebp pop edi pop esi pop ebx emms ret ;------------------------------------------------------------------------------ ; ; void yuv_to_yv12_mmx(uint8_t *y_out, ; uint8_t *u_out, ; uint8_t *v_out, ; uint8_t *src, ; int width, int height, int stride); ; ; Attention: This code assumes that width is a multiple of 16 ; ;------------------------------------------------------------------------------ cglobal yuv_to_yv12_mmx yuv_to_yv12_mmx: push ebx push esi push edi push ebp ; local vars allocation %define localsize 4 %define remainder esp sub esp, localsize ; function code mov eax, [esp + 40 + localsize] ; height -> eax mov ebx, [esp + 44 + localsize] ; stride -> ebx mov esi, [esp + 32 + localsize] ; src -> esi mov edi, [esp + 20 + localsize] ; y_out -> edi mov ecx, [esp + 36 + localsize] ; width -> ecx sub ebx, ecx ; stride - width -> ebx mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; mainder -> ebp shr ebp, 4 ; 16 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .y_inner_loop: movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movq [edi], mm1 ; write to y_out movq [edi + 8], mm2 movq [edi + 16], mm3 movq [edi + 24], mm4 movq [edi + 32], mm5 movq [edi + 40], mm6 movq [edi + 48], mm7 movq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .y_inner_loop dec ebp jz .y_outer_loop .y_remainder_loop: movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq [edi], mm1 ; write to y_out movq [edi + 8], mm2 add esi, 16 add edi, 16 dec ebp jnz .y_remainder_loop .y_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz near .y_inner_loop mov eax, [esp + 40 + localsize] ; height -> eax mov ebx, [esp + 44 + localsize] ; stride -> ebx mov ecx, [esp + 36 + localsize] ; width -> ecx mov edi, [esp + 24 + localsize] ; u_out -> edi shr ecx, 1 ; width / 2 -> ecx shr ebx, 1 ; stride / 2 -> ebx shr eax, 1 ; height / 2 -> eax sub ebx, ecx ; stride / 2 - width / 2 -> ebx mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; remainder -> ebp shr ebp, 3 ; 8 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .u_inner_loop: movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movq [edi], mm1 ; write to u_out movq [edi + 8], mm2 movq [edi + 16], mm3 movq [edi + 24], mm4 movq [edi + 32], mm5 movq [edi + 40], mm6 movq [edi + 48], mm7 movq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .u_inner_loop dec ebp jz .u_outer_loop .u_remainder_loop: movq mm1, [esi] ; read from src movq [edi], mm1 ; write to y_out add esi, 8 add edi, 8 dec ebp jnz .u_remainder_loop .u_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz .u_inner_loop mov eax, [esp + 40 + localsize] ; height -> eax mov ecx, [esp + 36 + localsize] ; width -> ecx mov edi, [esp + 28 + localsize] ; v_out -> edi shr ecx, 1 ; width / 2 -> ecx shr eax, 1 ; height / 2 -> eax mov edx, ecx mov ebp, ecx shr edx, 6 mov ecx, edx ; 64 bytes copied per iteration shl edx, 6 sub ebp, edx ; remainder -> ebp shr ebp, 3 ; 8 bytes per iteration add ebp, 1 mov [remainder], ebp mov edx, ecx .v_inner_loop: movq mm1, [esi] ; read from src movq mm2, [esi + 8] movq mm3, [esi + 16] movq mm4, [esi + 24] movq mm5, [esi + 32] movq mm6, [esi + 40] movq mm7, [esi + 48] movq mm0, [esi + 56] movq [edi], mm1 ; write to u_out movq [edi + 8], mm2 movq [edi + 16], mm3 movq [edi + 24], mm4 movq [edi + 32], mm5 movq [edi + 40], mm6 movq [edi + 48], mm7 movq [edi + 56], mm0 add esi, 64 add edi, 64 dec ecx jnz .v_inner_loop dec ebp jz .v_outer_loop .v_remainder_loop: movq mm1, [esi] ; read from src movq [edi], mm1 ; write to y_out add esi, 8 add edi, 8 dec ebp jnz .v_remainder_loop .v_outer_loop: mov ebp, [remainder] mov ecx, edx add edi, ebx dec eax jnz .v_inner_loop ; local vars deallocation add esp, localsize %undef localsize %undef remainder pop ebp pop edi pop esi pop ebx emms ret