diff -r 000000000000 -r bde4ae8d615e os/ossrv/genericopenlibs/liboil/src/i386/mult8x8_i386.c --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/os/ossrv/genericopenlibs/liboil/src/i386/mult8x8_i386.c Fri Jun 15 03:10:57 2012 +0200 @@ -0,0 +1,145 @@ +/* + * LIBOIL - Library of Optimized Inner Loops + * Copyright (c) 2003,2004 David A. Schleef + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR + * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, + * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, + * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING + * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ +//Portions Copyright (c) 2008-2009 Nokia Corporation and/or its subsidiary(-ies). All rights reserved. + +#ifdef HAVE_CONFIG_H +#include "config.h" +#endif + +#include +#include "liboil/simdpack/simdpack.h" + +#include + + +static void +mult8x8_s16_mmx(int16_t *dest, int16_t *src1, int16_t *src2, int dstr, int sstr1, + int sstr2) +{ +#if !defined(__WINSCW__) && !defined(__WINS__) + asm volatile( + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + + " add %3, %0 \n" + " add %4, %1 \n" + " add %5, %2 \n" + + " movq 0(%1), %%mm0 \n" + " pmullw 0(%2), %%mm0 \n" + " movq %%mm0, 0(%0) \n" + " movq 8(%1), %%mm1 \n" + " pmullw 8(%2), %%mm1 \n" + " movq %%mm1, 8(%0) \n" + " emms\n " + + : "+r" (dest), "+r" (src1), "+r" (src2) + : "m" (dstr), "m" (sstr1), "m" (sstr2)); +#endif +} + +OIL_DEFINE_IMPL_FULL (mult8x8_s16_mmx, mult8x8_s16, OIL_IMPL_FLAG_MMX); + + + +#ifdef __SYMBIAN32__ + +OilFunctionImpl* __oil_function_impl_mult8x8_s16_mmx, mult8x8_s16() { + return &_oil_function_impl_mult8x8_s16_mmx, mult8x8_s16; +} +#endif +