From patchwork Mon Oct 10 08:09:12 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: =?utf-8?b?5ZGo5pmT5YuH?= X-Patchwork-Id: 948 Delivered-To: ffmpegpatchwork@gmail.com Received: by 10.103.140.66 with SMTP id o63csp47176vsd; Mon, 10 Oct 2016 10:19:02 -0700 (PDT) X-Received: by 10.194.109.229 with SMTP id hv5mr30782650wjb.131.1476119942237; Mon, 10 Oct 2016 10:19:02 -0700 (PDT) Return-Path: Received: from ffbox0-bg.mplayerhq.hu (ffbox0-bg.ffmpeg.org. [79.124.17.100]) by mx.google.com with ESMTP id 18si11881244wjz.280.2016.10.10.10.19.01; Mon, 10 Oct 2016 10:19:02 -0700 (PDT) Received-SPF: pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) client-ip=79.124.17.100; Authentication-Results: mx.google.com; dkim=neutral (body hash did not verify) header.i=@126.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org; dmarc=fail (p=NONE dis=NONE) header.from=126.com Received: from [127.0.1.1] (localhost [127.0.0.1]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 01400689934; Mon, 10 Oct 2016 20:18:41 +0300 (EEST) X-Original-To: ffmpeg-devel@ffmpeg.org Delivered-To: ffmpeg-devel@ffmpeg.org Received: from m15-56.126.com (m15-56.126.com [220.181.15.56]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 3414D689B28 for ; Mon, 10 Oct 2016 11:08:58 +0300 (EEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=126.com; s=s110527; h=Date:From:Subject:MIME-Version:Message-ID; bh=6kX9h AgSzeKfc/SXmkDx8TutrDPfE6HCbr9LGze0fl0=; b=bQXNubAcCkP0zvWxgTVr7 Oa2nGa48zGBq15i7epOkK7JW93PCFLfbsw3FVXUsobS/AmHmuiMcYGMzywNh2FYI NaA/4187G/XGnihDKynTge/oVwaznOef9wrE4DUdpNdbrXmQEGasjC8d7EjVPP1z vav4kKGLS0TR2GCaHzaQjQ= Received: from ipfootball$126.com ( [114.242.206.180] ) by ajax-webmail-wmsvr56 (Coremail) ; Mon, 10 Oct 2016 16:09:12 +0800 (CST) X-Originating-IP: [114.242.206.180] Date: Mon, 10 Oct 2016 16:09:12 +0800 (CST) From: =?GBK?B?1tzP/tPC?= To: "ffmpeg-devel@ffmpeg.org" X-Priority: 3 X-Mailer: Coremail Webmail Server Version SP_ntes V3.5 build 20160729(86883.8884) Copyright (c) 2002-2016 www.mailtech.cn 126com MIME-Version: 1.0 Message-ID: <47f24766.8e00.157ada373f6.Coremail.ipfootball@126.com> X-Coremail-Locale: zh_CN X-CM-TRANSID: OMqowADH1v6pTPtX2MlMAA--.2198W X-CM-SenderInfo: 5lsi00hwedzza6rslhhfrp/1tbiVw7+E1Ph+hghJwABsj X-Coremail-Antispam: 1U5529EdanIXcx71UUUUU7vcSsGvfC2KfnxnUU== X-Mailman-Approved-At: Mon, 10 Oct 2016 20:18:39 +0300 X-Content-Filtered-By: Mailman/MimeDel 2.1.20 Subject: [FFmpeg-devel] [PATCH 3/7] avcodec/mips: loongson optimize mmi load and store operators X-BeenThere: ffmpeg-devel@ffmpeg.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: FFmpeg development discussions and patches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Reply-To: FFmpeg development discussions and patches Cc: "michael@niedermayer.cc" Errors-To: ffmpeg-devel-bounces@ffmpeg.org Sender: "ffmpeg-devel" From ec6fd41adff8541180f4e43e019e31ff06867789 Mon Sep 17 00:00:00 2001 From: Zhou Xiaoyong Date: Mon, 10 Oct 2016 14:31:39 +0800 Subject: [PATCH 3/7] avcodec/mips: loongson optimize mmi load and store operators 1.MMI_ load/store macros are defined in libavutil/mips/mmiutils.h 2.Replace some unnecessary unaligned access with aligned operator 3.The MMI_ load/store is compatible with cpu loongson2e/2f which not support instructions start with gs --- libavcodec/mips/blockdsp_mmi.c | 139 +-- libavcodec/mips/h264chroma_mmi.c | 193 ++-- libavcodec/mips/h264dsp_mmi.c | 2126 ++++++++++++++++++------------------- libavcodec/mips/h264pred_mmi.c | 259 +++-- libavcodec/mips/h264qpel_mmi.c | 777 ++++++-------- libavcodec/mips/hpeldsp_mmi.c | 549 +++++----- libavcodec/mips/idctdsp_mmi.c | 130 ++- libavcodec/mips/mpegvideo_mmi.c | 112 +- libavcodec/mips/pixblockdsp_mmi.c | 87 +- 9 files changed, 2066 insertions(+), 2306 deletions(-) diff --git a/libavcodec/mips/blockdsp_mmi.c b/libavcodec/mips/blockdsp_mmi.c index 6eb2bd7..1035dbb 100644 --- a/libavcodec/mips/blockdsp_mmi.c +++ b/libavcodec/mips/blockdsp_mmi.c @@ -22,11 +22,12 @@ */ #include "blockdsp_mips.h" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_fill_block16_mmi(uint8_t *block, uint8_t value, int line_size, int h) { double ftmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "mtc1 %[value], %[ftmp0] \n\t" @@ -34,15 +35,14 @@ void ff_fill_block16_mmi(uint8_t *block, uint8_t value, int line_size, int h) "punpcklbh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "1: \n\t" - "gssdlc1 %[ftmp0], 0x07(%[block]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[block]) \n\t" - PTR_ADDI "%[h], %[h], -0x01 \n\t" - "gssdlc1 %[ftmp0], 0x0f(%[block]) \n\t" - "gssdrc1 %[ftmp0], 0x08(%[block]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + PTR_ADDI "%[h], %[h], -0x01 \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x08) PTR_ADDU "%[block], %[block], %[line_size] \n\t" "bnez %[h], 1b \n\t" - : [block]"+&r"(block), [h]"+&r"(h), - [ftmp0]"=&f"(ftmp[0]) + : [ftmp0]"=&f"(ftmp[0]), + RESTRICT_ASM_ALL64 + [block]"+&r"(block), [h]"+&r"(h) : [value]"r"(value), [line_size]"r"((mips_reg)line_size) : "memory" ); @@ -51,6 +51,7 @@ void ff_fill_block16_mmi(uint8_t *block, uint8_t value, int line_size, int h) void ff_fill_block8_mmi(uint8_t *block, uint8_t value, int line_size, int h) { double ftmp0; + DECLARE_VAR_ALL64; __asm__ volatile ( "mtc1 %[value], %[ftmp0] \n\t" @@ -58,13 +59,13 @@ void ff_fill_block8_mmi(uint8_t *block, uint8_t value, int line_size, int h) "punpcklbh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "1: \n\t" - "gssdlc1 %[ftmp0], 0x07(%[block]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[block]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) PTR_ADDI "%[h], %[h], -0x01 \n\t" PTR_ADDU "%[block], %[block], %[line_size] \n\t" "bnez %[h], 1b \n\t" - : [block]"+&r"(block), [h]"+&r"(h), - [ftmp0]"=&f"(ftmp0) + : [ftmp0]"=&f"(ftmp0), + RESTRICT_ASM_ALL64 + [block]"+&r"(block), [h]"+&r"(h) : [value]"r"(value), [line_size]"r"((mips_reg)line_size) : "memory" ); @@ -77,14 +78,14 @@ void ff_clear_block_mmi(int16_t *block) __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp1] \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x00(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x10(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x20(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x30(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x40(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x50(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x60(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x70(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x00) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x10) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x20) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x30) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x40) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x50) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x60) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x70) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]) : [block]"r"(block) : "memory" @@ -98,61 +99,61 @@ void ff_clear_blocks_mmi(int16_t *block) __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp1] \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x00(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x10(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x20(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x30(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x40(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x50(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x60(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x70(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x00) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x10) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x20) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x30) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x40) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x50) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x60) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x70) - "gssqc1 %[ftmp0], %[ftmp1], 0x80(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x90(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xa0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xb0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xc0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xd0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xe0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0xf0(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x80) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x90) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xa0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xb0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xc0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xd0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xe0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0xf0) - "gssqc1 %[ftmp0], %[ftmp1], 0x100(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x110(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x120(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x130(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x140(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x150(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x160(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x170(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x100) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x110) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x120) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x130) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x140) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x150) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x160) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x170) - "gssqc1 %[ftmp0], %[ftmp1], 0x180(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x190(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1a0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1b0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1c0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1d0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1e0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x1f0(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x180) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x190) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1a0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1b0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1c0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1d0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1e0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x1f0) - "gssqc1 %[ftmp0], %[ftmp1], 0x200(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x210(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x220(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x230(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x240(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x250(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x260(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x270(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x200) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x210) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x220) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x230) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x240) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x250) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x260) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x270) - "gssqc1 %[ftmp0], %[ftmp1], 0x280(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x290(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2a0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2b0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2c0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2d0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2e0(%[block]) \n\t" - "gssqc1 %[ftmp0], %[ftmp1], 0x2f0(%[block]) \n\t" + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x280) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x290) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2a0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2b0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2c0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2d0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2e0) + MMI_SQC1(%[ftmp0], %[ftmp1], %[block], 0x2f0) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]) - : [block]"r"((mips_reg)block) + : [block]"r"((uint64_t *)block) : "memory" ); } diff --git a/libavcodec/mips/h264chroma_mmi.c b/libavcodec/mips/h264chroma_mmi.c index 3dd123d..417b4a2 100644 --- a/libavcodec/mips/h264chroma_mmi.c +++ b/libavcodec/mips/h264chroma_mmi.c @@ -24,7 +24,7 @@ #include "h264chroma_mips.h" #include "constants.h" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, int h, int x, int y) @@ -37,6 +37,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, double ftmp[10]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; if (D) { __asm__ volatile ( @@ -47,16 +48,13 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "mtc1 %[tmp0], %[ftmp9] \n\t" "pshufh %[C], %[C], %[ftmp0] \n\t" "pshufh %[D], %[D], %[ftmp0] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[stride] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp2], 0x01(%[src]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp4], 0x08(%[addr0]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp2], %[src], 0x01) + MMI_ULDC1(%[ftmp3], %[addr0], 0x00) + MMI_ULDC1(%[ftmp4], %[addr0], 0x01) "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" @@ -88,7 +86,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -98,6 +96,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) @@ -115,12 +114,11 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "pshufh %[A], %[A], %[ftmp0] \n\t" "pshufh %[E], %[E], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp7] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[step] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp2], %[addr0], 0x00) "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" @@ -139,7 +137,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -148,6 +146,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) @@ -162,9 +161,9 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "dli %[tmp0], 0x06 \n\t" "pshufh %[A], %[A], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp4] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" @@ -175,11 +174,10 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" @@ -190,7 +188,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x02 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" @@ -199,6 +197,7 @@ void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), @@ -219,6 +218,7 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, double ftmp[10]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; if (D) { __asm__ volatile ( @@ -229,16 +229,13 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "mtc1 %[tmp0], %[ftmp9] \n\t" "pshufh %[C], %[C], %[ftmp0] \n\t" "pshufh %[D], %[D], %[ftmp0] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[stride] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp2], 0x01(%[src]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp4], 0x08(%[addr0]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp2], %[src], 0x01) + MMI_ULDC1(%[ftmp3], %[addr0], 0x00) + MMI_ULDC1(%[ftmp4], %[addr0], 0x01) "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" @@ -269,10 +266,10 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t" "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -282,6 +279,7 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) @@ -299,12 +297,11 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "pshufh %[A], %[A], %[ftmp0] \n\t" "pshufh %[E], %[E], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp7] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[step] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp2], %[addr0], 0x00) "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" @@ -322,10 +319,10 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -334,6 +331,7 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) @@ -348,9 +346,9 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "dli %[tmp0], 0x06 \n\t" "pshufh %[A], %[A], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp4] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" @@ -360,14 +358,13 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t" "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" @@ -377,10 +374,10 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp4] \n\t" "psrlh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x02 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" @@ -389,6 +386,7 @@ void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), @@ -409,7 +407,7 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, double ftmp[8]; uint64_t tmp[1]; mips_reg addr[1]; - uint64_t low32; + DECLARE_VAR_LOW32; if (D) { __asm__ volatile ( @@ -420,16 +418,13 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "mtc1 %[tmp0], %[ftmp7] \n\t" "pshufh %[C], %[C], %[ftmp0] \n\t" "pshufh %[D], %[D], %[ftmp0] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[stride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) + MMI_ULWC1(%[ftmp2], %[src], 0x01) + MMI_ULWC1(%[ftmp3], %[addr0], 0x00) + MMI_ULWC1(%[ftmp4], %[addr0], 0x01) "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" @@ -448,7 +443,7 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "addi %[h], %[h], -0x01 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -457,10 +452,10 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), [A]"f"(A), [B]"f"(B), [C]"f"(C), [D]"f"(D) @@ -475,12 +470,11 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "pshufh %[A], %[A], %[ftmp0] \n\t" "pshufh %[E], %[E], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp5] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[step] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) + MMI_ULWC1(%[ftmp2], %[addr0], 0x00) "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" @@ -492,7 +486,7 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "addi %[h], %[h], -0x01 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -500,10 +494,10 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step), [ff_pw_32]"f"(ff_pw_32), [A]"f"(A), [E]"f"(E) @@ -515,27 +509,26 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "dli %[tmp0], 0x06 \n\t" "pshufh %[A], %[A], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp3] \n\t" + "1: \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "addi %[h], %[h], -0x02 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" @@ -543,9 +536,9 @@ void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), [A]"f"(A) : "memory" @@ -564,7 +557,7 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, double ftmp[8]; uint64_t tmp[1]; mips_reg addr[1]; - uint64_t low32; + DECLARE_VAR_LOW32; if (D) { __asm__ volatile ( @@ -575,16 +568,13 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "mtc1 %[tmp0], %[ftmp7] \n\t" "pshufh %[C], %[C], %[ftmp0] \n\t" "pshufh %[D], %[D], %[ftmp0] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src], %[stride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) + MMI_ULWC1(%[ftmp2], %[src], 0x01) + MMI_ULWC1(%[ftmp3], %[addr0], 0x00) + MMI_ULWC1(%[ftmp4], %[addr0], 0x01) "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" @@ -602,10 +592,10 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -614,10 +604,10 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), [A]"f"(A), [B]"f"(B), [C]"f"(C), [D]"f"(D) @@ -634,10 +624,8 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "mtc1 %[tmp0], %[ftmp5] \n\t" "1: \n\t" PTR_ADDU "%[addr0], %[src], %[step] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) + MMI_ULWC1(%[ftmp2], %[addr0], 0x00) "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" @@ -648,10 +636,10 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x01 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" "bnez %[h], 1b \n\t" @@ -659,10 +647,10 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [addr0]"=&r"(addr[0]), [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step), [ff_pw_32]"f"(ff_pw_32), [A]"f"(A), [E]"f"(E) @@ -674,31 +662,30 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, "dli %[tmp0], 0x06 \n\t" "pshufh %[A], %[A], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp3] \n\t" + "1: \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "pmullh %[ftmp1], %[ftmp2], %[A] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "addi %[h], %[h], -0x02 \n\t" - "swc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" @@ -706,9 +693,9 @@ void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, uint8_t *src, int stride, : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32), [A]"f"(A) : "memory" diff --git a/libavcodec/mips/h264dsp_mmi.c b/libavcodec/mips/h264dsp_mmi.c index a550eee..ac6fa99 100644 --- a/libavcodec/mips/h264dsp_mmi.c +++ b/libavcodec/mips/h264dsp_mmi.c @@ -25,27 +25,24 @@ #include "libavcodec/bit_depth_template.c" #include "h264dsp_mips.h" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_h264_add_pixels4_8_mmi(uint8_t *dst, int16_t *src, int stride) { double ftmp[9]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ALL64; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "ldc1 %[ftmp1], 0x00(%[src]) \n\t" - "ldc1 %[ftmp2], 0x08(%[src]) \n\t" - "ldc1 %[ftmp3], 0x10(%[src]) \n\t" - "ldc1 %[ftmp4], 0x18(%[src]) \n\t" - "uld %[low32], 0x00(%[dst0]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" - "uld %[low32], 0x00(%[dst1]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" - "uld %[low32], 0x00(%[dst2]) \n\t" - "mtc1 %[low32], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[dst3]) \n\t" - "mtc1 %[low32], %[ftmp8] \n\t" + MMI_LDC1(%[ftmp1], %[src], 0x00) + MMI_LDC1(%[ftmp2], %[src], 0x08) + MMI_LDC1(%[ftmp3], %[src], 0x10) + MMI_LDC1(%[ftmp4], %[src], 0x18) + MMI_ULWC1(%[ftmp5], %[dst0], 0x00) + MMI_ULWC1(%[ftmp6], %[dst1], 0x00) + MMI_ULWC1(%[ftmp7], %[dst2], 0x00) + MMI_ULWC1(%[ftmp8], %[dst3], 0x00) "punpcklbh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" @@ -58,20 +55,17 @@ void ff_h264_add_pixels4_8_mmi(uint8_t *dst, int16_t *src, int stride) "packushb %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "packushb %[ftmp3], %[ftmp3], %[ftmp0] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "gsswlc1 %[ftmp1], 0x03(%[dst0]) \n\t" - "gsswrc1 %[ftmp1], 0x00(%[dst0]) \n\t" - "gsswlc1 %[ftmp2], 0x03(%[dst1]) \n\t" - "gsswrc1 %[ftmp2], 0x00(%[dst1]) \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst2]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst2]) \n\t" - "gsswlc1 %[ftmp4], 0x03(%[dst3]) \n\t" - "gsswrc1 %[ftmp4], 0x00(%[dst3]) \n\t" + MMI_SWC1(%[ftmp1], %[dst0], 0x00) + MMI_SWC1(%[ftmp2], %[dst1], 0x00) + MMI_SWC1(%[ftmp3], %[dst2], 0x00) + MMI_SWC1(%[ftmp4], %[dst3], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), - [ftmp8]"=&f"(ftmp[8]), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ALL64 + [ftmp8]"=&f"(ftmp[8]) : [dst0]"r"(dst), [dst1]"r"(dst+stride), [dst2]"r"(dst+2*stride), [dst3]"r"(dst+3*stride), [src]"r"(src) @@ -85,18 +79,20 @@ void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride) { double ftmp[12]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( "dli %[tmp0], 0x01 \n\t" - "ldc1 %[ftmp0], 0x00(%[block]) \n\t" + MMI_LDC1(%[ftmp0], %[block], 0x00) "mtc1 %[tmp0], %[ftmp8] \n\t" - "ldc1 %[ftmp1], 0x08(%[block]) \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x08) "dli %[tmp0], 0x06 \n\t" - "ldc1 %[ftmp2], 0x10(%[block]) \n\t" + MMI_LDC1(%[ftmp2], %[block], 0x10) "mtc1 %[tmp0], %[ftmp9] \n\t" "psrah %[ftmp4], %[ftmp1], %[ftmp8] \n\t" - "ldc1 %[ftmp3], 0x18(%[block]) \n\t" + MMI_LDC1(%[ftmp3], %[block], 0x18) "psrah %[ftmp5], %[ftmp3], %[ftmp8] \n\t" "psubh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" "paddh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" @@ -126,14 +122,13 @@ void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "paddh %[ftmp11], %[ftmp4], %[ftmp5] \n\t" "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" "psubh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" - "sdc1 %[ftmp7], 0x00(%[block]) \n\t" - "sdc1 %[ftmp7], 0x08(%[block]) \n\t" - "sdc1 %[ftmp7], 0x10(%[block]) \n\t" - "sdc1 %[ftmp7], 0x18(%[block]) \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp7], %[block], 0x00) + MMI_SDC1(%[ftmp7], %[block], 0x08) + MMI_SDC1(%[ftmp7], %[block], 0x10) + MMI_SDC1(%[ftmp7], %[block], 0x18) + MMI_ULWC1(%[ftmp2], %[dst], 0x00) "psrah %[ftmp3], %[ftmp10], %[ftmp9] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) "psrah %[ftmp4], %[ftmp11], %[ftmp9] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -141,33 +136,32 @@ void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "packushb %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "gsswlc1 %[ftmp2], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp2], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" + MMI_SWC1(%[ftmp2], %[dst], 0x00) + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) PTR_ADDU "%[dst], %[dst], %[stride] \n\t" PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp2], %[dst], 0x00) "psrah %[ftmp5], %[ftmp5], %[ftmp9] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" "paddh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" "paddh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" "packushb %[ftmp2], %[ftmp2], %[ftmp7] \n\t" - "gsswlc1 %[ftmp2], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp2], %[dst], 0x00) "packushb %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [ftmp11]"=&f"(ftmp[11]), - [tmp0]"=&r"(tmp[0]), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT + [tmp0]"=&r"(tmp[0]) : [dst]"r"(dst), [block]"r"(block), [stride]"r"((mips_reg)stride), [ff_pw_32]"f"(ff_pw_32) : "memory" @@ -179,464 +173,450 @@ void ff_h264_idct_add_8_mmi(uint8_t *dst, int16_t *block, int stride) void ff_h264_idct8_add_8_mmi(uint8_t *dst, int16_t *block, int stride) { double ftmp[16]; - uint64_t tmp[8]; + uint64_t tmp[7]; mips_reg addr[1]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( - "lhu %[tmp0], 0x00(%[block]) \n\t" - PTR_ADDI "$29, $29, -0x20 \n\t" - PTR_ADDIU "%[tmp0], %[tmp0], 0x20 \n\t" - "ldc1 %[ftmp1], 0x10(%[block]) \n\t" - "sh %[tmp0], 0x00(%[block]) \n\t" - "ldc1 %[ftmp2], 0x20(%[block]) \n\t" - "dli %[tmp0], 0x01 \n\t" - "ldc1 %[ftmp3], 0x30(%[block]) \n\t" - "mtc1 %[tmp0], %[ftmp8] \n\t" - "ldc1 %[ftmp5], 0x50(%[block]) \n\t" - "ldc1 %[ftmp6], 0x60(%[block]) \n\t" - "ldc1 %[ftmp7], 0x70(%[block]) \n\t" - "mov.d %[ftmp0], %[ftmp1] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" - "psrah %[ftmp4], %[ftmp5], %[ftmp8] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp3] \n\t" - "psrah %[ftmp3], %[ftmp3], %[ftmp8] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "psrah %[ftmp7], %[ftmp7], %[ftmp8] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "dli %[tmp0], 0x02 \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "mtc1 %[tmp0], %[ftmp9] \n\t" - "mov.d %[ftmp7], %[ftmp1] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" - "psrah %[ftmp3], %[ftmp4], %[ftmp9] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "psrah %[ftmp0], %[ftmp0], %[ftmp9] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp9] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "mov.d %[ftmp5], %[ftmp6] \n\t" - "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "psrah %[ftmp4], %[ftmp2], %[ftmp8] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" - "ldc1 %[ftmp2], 0x00(%[block]) \n\t" - "ldc1 %[ftmp5], 0x40(%[block]) \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp2] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "psubh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp2] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp6] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" - "psubh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp7] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp4] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "psubh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" - "sdc1 %[ftmp6], 0x00(%[block]) \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" - "punpckhhw %[ftmp6], %[ftmp7], %[ftmp0] \n\t" - "punpcklhw %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "punpckhhw %[ftmp0], %[ftmp3], %[ftmp1] \n\t" - "punpcklhw %[ftmp3], %[ftmp3], %[ftmp1] \n\t" - "punpckhwd %[ftmp1], %[ftmp7], %[ftmp3] \n\t" - "punpcklwd %[ftmp7], %[ftmp7], %[ftmp3] \n\t" - "punpckhwd %[ftmp3], %[ftmp6], %[ftmp0] \n\t" - "punpcklwd %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "ldc1 %[ftmp0], 0x00(%[block]) \n\t" - "sdc1 %[ftmp7], 0x00($29) \n\t" - "sdc1 %[ftmp1], 0x10($29) \n\t" - "dmfc1 %[tmp1], %[ftmp6] \n\t" - "dmfc1 %[tmp3], %[ftmp3] \n\t" - "punpckhhw %[ftmp3], %[ftmp5], %[ftmp2] \n\t" - "punpcklhw %[ftmp5], %[ftmp5], %[ftmp2] \n\t" - "punpckhhw %[ftmp2], %[ftmp4], %[ftmp0] \n\t" - "punpcklhw %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "punpckhwd %[ftmp0], %[ftmp5], %[ftmp4] \n\t" - "punpcklwd %[ftmp5], %[ftmp5], %[ftmp4] \n\t" - "punpckhwd %[ftmp4], %[ftmp3], %[ftmp2] \n\t" - "punpcklwd %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "sdc1 %[ftmp5], 0x08($29) \n\t" - "sdc1 %[ftmp0], 0x18($29) \n\t" - "dmfc1 %[tmp2], %[ftmp3] \n\t" - "dmfc1 %[tmp4], %[ftmp4] \n\t" - "ldc1 %[ftmp1], 0x18(%[block]) \n\t" - "ldc1 %[ftmp6], 0x28(%[block]) \n\t" - "ldc1 %[ftmp2], 0x38(%[block]) \n\t" - "ldc1 %[ftmp0], 0x58(%[block]) \n\t" - "ldc1 %[ftmp3], 0x68(%[block]) \n\t" - "ldc1 %[ftmp4], 0x78(%[block]) \n\t" - "mov.d %[ftmp7], %[ftmp1] \n\t" - "psrah %[ftmp5], %[ftmp0], %[ftmp8] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "psrah %[ftmp2], %[ftmp2], %[ftmp8] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp8] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "mov.d %[ftmp4], %[ftmp1] \n\t" - "psrah %[ftmp2], %[ftmp5], %[ftmp9] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" - "psrah %[ftmp7], %[ftmp7], %[ftmp9] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "psrah %[ftmp0], %[ftmp0], %[ftmp9] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "mov.d %[ftmp0], %[ftmp3] \n\t" - "psrah %[ftmp3], %[ftmp3], %[ftmp8] \n\t" - "psrah %[ftmp5], %[ftmp6], %[ftmp8] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" - "ldc1 %[ftmp6], 0x08(%[block]) \n\t" - "ldc1 %[ftmp0], 0x48(%[block]) \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp6] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp6] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "sdc1 %[ftmp3], 0x08(%[block]) \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "punpckhhw %[ftmp3], %[ftmp4], %[ftmp7] \n\t" - "punpcklhw %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "punpckhhw %[ftmp7], %[ftmp2], %[ftmp1] \n\t" - "punpcklhw %[ftmp2], %[ftmp2], %[ftmp1] \n\t" - "punpckhwd %[ftmp1], %[ftmp4], %[ftmp2] \n\t" - "punpcklwd %[ftmp4], %[ftmp4], %[ftmp2] \n\t" - "punpckhwd %[ftmp2], %[ftmp3], %[ftmp7] \n\t" - "punpcklwd %[ftmp3], %[ftmp3], %[ftmp7] \n\t" - "ldc1 %[ftmp7], 0x08(%[block]) \n\t" - "dmfc1 %[tmp5], %[ftmp4] \n\t" - "dmfc1 %[tmp7], %[ftmp1] \n\t" - "mov.d %[ftmp12], %[ftmp3] \n\t" - "mov.d %[ftmp14], %[ftmp2] \n\t" - "punpckhhw %[ftmp2], %[ftmp0], %[ftmp6] \n\t" - "punpcklhw %[ftmp0], %[ftmp0], %[ftmp6] \n\t" - "punpckhhw %[ftmp6], %[ftmp5], %[ftmp7] \n\t" - "punpcklhw %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "punpckhwd %[ftmp7], %[ftmp0], %[ftmp5] \n\t" - "punpcklwd %[ftmp0], %[ftmp0], %[ftmp5] \n\t" - "punpckhwd %[ftmp5], %[ftmp2], %[ftmp6] \n\t" - "punpcklwd %[ftmp2], %[ftmp2], %[ftmp6] \n\t" - "dmfc1 %[tmp6], %[ftmp0] \n\t" - "mov.d %[ftmp11], %[ftmp7] \n\t" - "mov.d %[ftmp13], %[ftmp2] \n\t" - "mov.d %[ftmp15], %[ftmp5] \n\t" - PTR_ADDIU "%[addr0], %[dst], 0x04 \n\t" - "dmtc1 %[tmp7], %[ftmp7] \n\t" - "dmtc1 %[tmp3], %[ftmp6] \n\t" - "ldc1 %[ftmp1], 0x10($29) \n\t" - "dmtc1 %[tmp1], %[ftmp3] \n\t" - "mov.d %[ftmp4], %[ftmp1] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" - "psrah %[ftmp0], %[ftmp7], %[ftmp8] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp4] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp14] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp6] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" - "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp14] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp14] \n\t" - "psrah %[ftmp5], %[ftmp14], %[ftmp8] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "mov.d %[ftmp5], %[ftmp1] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" - "psrah %[ftmp6], %[ftmp0], %[ftmp9] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp9] \n\t" - "psrah %[ftmp7], %[ftmp7], %[ftmp9] \n\t" - "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "mov.d %[ftmp7], %[ftmp12] \n\t" - "psrah %[ftmp2], %[ftmp12], %[ftmp8] \n\t" - "psrah %[ftmp0], %[ftmp3], %[ftmp8] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "ldc1 %[ftmp3], 0x00($29) \n\t" - "dmtc1 %[tmp5], %[ftmp7] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp2] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "psubh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "sdc1 %[ftmp3], 0x00($29) \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" - "sdc1 %[ftmp0], 0x10($29) \n\t" - "dmfc1 %[tmp1], %[ftmp2] \n\t" - "xor %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "sdc1 %[ftmp2], 0x00(%[block]) \n\t" - "sdc1 %[ftmp2], 0x08(%[block]) \n\t" - "sdc1 %[ftmp2], 0x10(%[block]) \n\t" - "sdc1 %[ftmp2], 0x18(%[block]) \n\t" - "sdc1 %[ftmp2], 0x20(%[block]) \n\t" - "sdc1 %[ftmp2], 0x28(%[block]) \n\t" - "sdc1 %[ftmp2], 0x30(%[block]) \n\t" - "sdc1 %[ftmp2], 0x38(%[block]) \n\t" - "sdc1 %[ftmp2], 0x40(%[block]) \n\t" - "sdc1 %[ftmp2], 0x48(%[block]) \n\t" - "sdc1 %[ftmp2], 0x50(%[block]) \n\t" - "sdc1 %[ftmp2], 0x58(%[block]) \n\t" - "sdc1 %[ftmp2], 0x60(%[block]) \n\t" - "sdc1 %[ftmp2], 0x68(%[block]) \n\t" - "sdc1 %[ftmp2], 0x70(%[block]) \n\t" - "sdc1 %[ftmp2], 0x78(%[block]) \n\t" - "dli %[tmp3], 0x06 \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "mtc1 %[tmp3], %[ftmp10] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" - "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" - "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "psrah %[ftmp6], %[ftmp6], %[ftmp10] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" - "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "ldc1 %[ftmp5], 0x00($29) \n\t" - "ldc1 %[ftmp4], 0x10($29) \n\t" - "dmtc1 %[tmp1], %[ftmp6] \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "psrah %[ftmp7], %[ftmp7], %[ftmp10] \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" - "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp5] \n\t" - "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - PTR_ADDU "%[dst], %[dst], %[stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "gslwxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" - "psrah %[ftmp6], %[ftmp6], %[ftmp10] \n\t" - "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" - "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp0], 0x00(%[dst], %[stride]) \n\t" - "dmtc1 %[tmp4], %[ftmp1] \n\t" - "dmtc1 %[tmp2], %[ftmp6] \n\t" - "ldc1 %[ftmp4], 0x18($29) \n\t" - "mov.d %[ftmp5], %[ftmp4] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp8] \n\t" - "psrah %[ftmp7], %[ftmp11], %[ftmp8] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp11] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp15] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp11] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp1] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" - "psubh %[ftmp3], %[ftmp11], %[ftmp1] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp15] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp15] \n\t" - "psrah %[ftmp2], %[ftmp15], %[ftmp8] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "mov.d %[ftmp2], %[ftmp4] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp9] \n\t" - "psrah %[ftmp1], %[ftmp7], %[ftmp9] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp9] \n\t" - "psrah %[ftmp3], %[ftmp3], %[ftmp9] \n\t" - "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "psubh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" - "mov.d %[ftmp3], %[ftmp13] \n\t" - "psrah %[ftmp0], %[ftmp13], %[ftmp8] \n\t" - "psrah %[ftmp7], %[ftmp6], %[ftmp8] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" - "ldc1 %[ftmp6], 0x08($29) \n\t" - "dmtc1 %[tmp6], %[ftmp3] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp7] \n\t" - "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "paddh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "psubh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "paddh %[ftmp1], %[ftmp1], %[ftmp6] \n\t" - "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" - "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" - "sdc1 %[ftmp6], 0x08($29) \n\t" - "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "sdc1 %[ftmp7], 0x18($29) \n\t" - "dmfc1 %[tmp2], %[ftmp0] \n\t" - "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" - "gslwxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - "psrah %[ftmp2], %[ftmp2], %[ftmp10] \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" - "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" - "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "gsswlc1 %[ftmp6], 0x03(%[addr0]) \n\t" - "gsswrc1 %[ftmp6], 0x00(%[addr0]) \n\t" - "gsswxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" - "gslwxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" - "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" - "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" - "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "gsswlc1 %[ftmp6], 0x03(%[addr0]) \n\t" - "gsswrc1 %[ftmp6], 0x00(%[addr0]) \n\t" - "gsswxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - "ldc1 %[ftmp2], 0x08($29) \n\t" - "ldc1 %[ftmp5], 0x18($29) \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "dmtc1 %[tmp2], %[ftmp1] \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" - "gslwxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - "psrah %[ftmp3], %[ftmp3], %[ftmp10] \n\t" - "psrah %[ftmp2], %[ftmp2], %[ftmp10] \n\t" - "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" - "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "gsswlc1 %[ftmp6], 0x03(%[addr0]) \n\t" - "gsswrc1 %[ftmp6], 0x00(%[addr0]) \n\t" - "gsswxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" - "gslwxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" - "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" - "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "paddh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" - "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" - "gsswlc1 %[ftmp6], 0x03(%[addr0]) \n\t" - "gsswrc1 %[ftmp6], 0x00(%[addr0]) \n\t" - "gsswxc1 %[ftmp7], 0x00(%[addr0], %[stride]) \n\t" - PTR_ADDIU "$29, $29, 0x20 \n\t" + "lhu %[tmp0], 0x00(%[block]) \n\t" + PTR_ADDI "$29, $29, -0x20 \n\t" + PTR_ADDIU "%[tmp0], %[tmp0], 0x20 \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x10) + "sh %[tmp0], 0x00(%[block]) \n\t" + MMI_LDC1(%[ftmp2], %[block], 0x20) + "dli %[tmp0], 0x01 \n\t" + MMI_LDC1(%[ftmp3], %[block], 0x30) + "mtc1 %[tmp0], %[ftmp8] \n\t" + MMI_LDC1(%[ftmp5], %[block], 0x50) + MMI_LDC1(%[ftmp6], %[block], 0x60) + MMI_LDC1(%[ftmp7], %[block], 0x70) + "mov.d %[ftmp0], %[ftmp1] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" + "psrah %[ftmp4], %[ftmp5], %[ftmp8] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp3] \n\t" + "psrah %[ftmp3], %[ftmp3], %[ftmp8] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "psrah %[ftmp7], %[ftmp7], %[ftmp8] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" + "dli %[tmp0], 0x02 \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "mtc1 %[tmp0], %[ftmp9] \n\t" + "mov.d %[ftmp7], %[ftmp1] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" + "psrah %[ftmp3], %[ftmp4], %[ftmp9] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" + "psrah %[ftmp0], %[ftmp0], %[ftmp9] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" + "psrah %[ftmp5], %[ftmp5], %[ftmp9] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "mov.d %[ftmp5], %[ftmp6] \n\t" + "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" + "psrah %[ftmp4], %[ftmp2], %[ftmp8] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" + MMI_LDC1(%[ftmp2], %[block], 0x00) + MMI_LDC1(%[ftmp5], %[block], 0x40) + "paddh %[ftmp5], %[ftmp5], %[ftmp2] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" + "psubh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp2] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp6] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" + "psubh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp7] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp4] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" + "psubh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" + MMI_SDC1(%[ftmp6], %[block], 0x00) + "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" + "punpckhhw %[ftmp6], %[ftmp7], %[ftmp0] \n\t" + "punpcklhw %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + "punpckhhw %[ftmp0], %[ftmp3], %[ftmp1] \n\t" + "punpcklhw %[ftmp3], %[ftmp3], %[ftmp1] \n\t" + "punpckhwd %[ftmp1], %[ftmp7], %[ftmp3] \n\t" + "punpcklwd %[ftmp7], %[ftmp7], %[ftmp3] \n\t" + "punpckhwd %[ftmp3], %[ftmp6], %[ftmp0] \n\t" + "punpcklwd %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + MMI_LDC1(%[ftmp0], %[block], 0x00) + MMI_SDC1(%[ftmp7], $29, 0x00) + MMI_SDC1(%[ftmp1], $29, 0x10) + "dmfc1 %[tmp1], %[ftmp6] \n\t" + "dmfc1 %[tmp3], %[ftmp3] \n\t" + "punpckhhw %[ftmp3], %[ftmp5], %[ftmp2] \n\t" + "punpcklhw %[ftmp5], %[ftmp5], %[ftmp2] \n\t" + "punpckhhw %[ftmp2], %[ftmp4], %[ftmp0] \n\t" + "punpcklhw %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "punpckhwd %[ftmp0], %[ftmp5], %[ftmp4] \n\t" + "punpcklwd %[ftmp5], %[ftmp5], %[ftmp4] \n\t" + "punpckhwd %[ftmp4], %[ftmp3], %[ftmp2] \n\t" + "punpcklwd %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp5], $29, 0x08) + MMI_SDC1(%[ftmp0], $29, 0x18) + "dmfc1 %[tmp2], %[ftmp3] \n\t" + "dmfc1 %[tmp4], %[ftmp4] \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x18) + MMI_LDC1(%[ftmp6], %[block], 0x28) + MMI_LDC1(%[ftmp2], %[block], 0x38) + MMI_LDC1(%[ftmp0], %[block], 0x58) + MMI_LDC1(%[ftmp3], %[block], 0x68) + MMI_LDC1(%[ftmp4], %[block], 0x78) + "mov.d %[ftmp7], %[ftmp1] \n\t" + "psrah %[ftmp5], %[ftmp0], %[ftmp8] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "psrah %[ftmp2], %[ftmp2], %[ftmp8] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp8] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "mov.d %[ftmp4], %[ftmp1] \n\t" + "psrah %[ftmp2], %[ftmp5], %[ftmp9] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" + "psrah %[ftmp7], %[ftmp7], %[ftmp9] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" + "psrah %[ftmp0], %[ftmp0], %[ftmp9] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "mov.d %[ftmp0], %[ftmp3] \n\t" + "psrah %[ftmp3], %[ftmp3], %[ftmp8] \n\t" + "psrah %[ftmp5], %[ftmp6], %[ftmp8] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" + MMI_LDC1(%[ftmp6], %[block], 0x08) + MMI_LDC1(%[ftmp0], %[block], 0x48) + "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp6] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp5] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp6] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp3], %[block], 0x08) + "psubh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" + "punpckhhw %[ftmp3], %[ftmp4], %[ftmp7] \n\t" + "punpcklhw %[ftmp4], %[ftmp4], %[ftmp7] \n\t" + "punpckhhw %[ftmp7], %[ftmp2], %[ftmp1] \n\t" + "punpcklhw %[ftmp2], %[ftmp2], %[ftmp1] \n\t" + "punpckhwd %[ftmp1], %[ftmp4], %[ftmp2] \n\t" + "punpcklwd %[ftmp4], %[ftmp4], %[ftmp2] \n\t" + "punpckhwd %[ftmp2], %[ftmp3], %[ftmp7] \n\t" + "punpcklwd %[ftmp3], %[ftmp3], %[ftmp7] \n\t" + MMI_LDC1(%[ftmp7], %[block], 0x08) + "dmfc1 %[tmp5], %[ftmp4] \n\t" + "mov.d %[ftmp10], %[ftmp1] \n\t" + "mov.d %[ftmp12], %[ftmp3] \n\t" + "mov.d %[ftmp14], %[ftmp2] \n\t" + "punpckhhw %[ftmp2], %[ftmp0], %[ftmp6] \n\t" + "punpcklhw %[ftmp0], %[ftmp0], %[ftmp6] \n\t" + "punpckhhw %[ftmp6], %[ftmp5], %[ftmp7] \n\t" + "punpcklhw %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "punpckhwd %[ftmp7], %[ftmp0], %[ftmp5] \n\t" + "punpcklwd %[ftmp0], %[ftmp0], %[ftmp5] \n\t" + "punpckhwd %[ftmp5], %[ftmp2], %[ftmp6] \n\t" + "punpcklwd %[ftmp2], %[ftmp2], %[ftmp6] \n\t" + "dmfc1 %[tmp6], %[ftmp0] \n\t" + "mov.d %[ftmp11], %[ftmp7] \n\t" + "mov.d %[ftmp13], %[ftmp2] \n\t" + "mov.d %[ftmp15], %[ftmp5] \n\t" + PTR_ADDIU "%[addr0], %[dst], 0x04 \n\t" + "mov.d %[ftmp7], %[ftmp10] \n\t" + "dmtc1 %[tmp3], %[ftmp6] \n\t" + MMI_LDC1(%[ftmp1], $29, 0x10) + "dmtc1 %[tmp1], %[ftmp3] \n\t" + "mov.d %[ftmp4], %[ftmp1] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" + "psrah %[ftmp0], %[ftmp7], %[ftmp8] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp4] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp14] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp6] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" + "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp14] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp14] \n\t" + "psrah %[ftmp5], %[ftmp14], %[ftmp8] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "mov.d %[ftmp5], %[ftmp1] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp9] \n\t" + "psrah %[ftmp6], %[ftmp0], %[ftmp9] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp9] \n\t" + "psrah %[ftmp7], %[ftmp7], %[ftmp9] \n\t" + "psubh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "mov.d %[ftmp7], %[ftmp12] \n\t" + "psrah %[ftmp2], %[ftmp12], %[ftmp8] \n\t" + "psrah %[ftmp0], %[ftmp3], %[ftmp8] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" + MMI_LDC1(%[ftmp3], $29, 0x00) + "dmtc1 %[tmp5], %[ftmp7] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp2] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp2] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" + "psubh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" + MMI_SDC1(%[ftmp3], $29, 0x00) + "psubh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" + MMI_SDC1(%[ftmp0], $29, 0x10) + "dmfc1 %[tmp1], %[ftmp2] \n\t" + "xor %[ftmp2], %[ftmp2], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp2], %[block], 0x00) + MMI_SDC1(%[ftmp2], %[block], 0x08) + MMI_SDC1(%[ftmp2], %[block], 0x10) + MMI_SDC1(%[ftmp2], %[block], 0x18) + MMI_SDC1(%[ftmp2], %[block], 0x20) + MMI_SDC1(%[ftmp2], %[block], 0x28) + MMI_SDC1(%[ftmp2], %[block], 0x30) + MMI_SDC1(%[ftmp2], %[block], 0x38) + MMI_SDC1(%[ftmp2], %[block], 0x40) + MMI_SDC1(%[ftmp2], %[block], 0x48) + MMI_SDC1(%[ftmp2], %[block], 0x50) + MMI_SDC1(%[ftmp2], %[block], 0x58) + MMI_SDC1(%[ftmp2], %[block], 0x60) + MMI_SDC1(%[ftmp2], %[block], 0x68) + MMI_SDC1(%[ftmp2], %[block], 0x70) + MMI_SDC1(%[ftmp2], %[block], 0x78) + "dli %[tmp3], 0x06 \n\t" + "mtc1 %[tmp3], %[ftmp10] \n\t" + MMI_ULWC1(%[ftmp3], %[dst], 0x00) + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) + "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" + "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" + "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + MMI_SWC1(%[ftmp3], %[dst], 0x00) + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + MMI_ULWC1(%[ftmp3], %[dst], 0x00) + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) + "psrah %[ftmp6], %[ftmp6], %[ftmp10] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" + "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" + "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + MMI_SWC1(%[ftmp3], %[dst], 0x00) + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) + MMI_LDC1(%[ftmp5], $29, 0x00) + MMI_LDC1(%[ftmp4], $29, 0x10) + "dmtc1 %[tmp1], %[ftmp6] \n\t" + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + MMI_ULWC1(%[ftmp3], %[dst], 0x00) + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) + "psrah %[ftmp7], %[ftmp7], %[ftmp10] \n\t" + "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" + "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp5] \n\t" + "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + MMI_SWC1(%[ftmp3], %[dst], 0x00) + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + PTR_ADDU "%[dst], %[dst], %[stride] \n\t" + MMI_ULWC1(%[ftmp3], %[dst], 0x00) + MMI_LWXC1(%[ftmp0], %[dst], %[stride], 0x00) + "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" + "psrah %[ftmp6], %[ftmp6], %[ftmp10] \n\t" + "punpcklbh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "punpcklbh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" + "packushb %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "packushb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + MMI_SWC1(%[ftmp3], %[dst], 0x00) + MMI_SWXC1(%[ftmp0], %[dst], %[stride], 0x00) + "dmtc1 %[tmp4], %[ftmp1] \n\t" + "dmtc1 %[tmp2], %[ftmp6] \n\t" + MMI_LDC1(%[ftmp4], $29, 0x18) + "mov.d %[ftmp5], %[ftmp4] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp8] \n\t" + "psrah %[ftmp7], %[ftmp11], %[ftmp8] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp11] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp5] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp15] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp11] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp1] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" + "psubh %[ftmp3], %[ftmp11], %[ftmp1] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp15] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp15] \n\t" + "psrah %[ftmp2], %[ftmp15], %[ftmp8] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp1] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" + "mov.d %[ftmp2], %[ftmp4] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp9] \n\t" + "psrah %[ftmp1], %[ftmp7], %[ftmp9] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" + "psrah %[ftmp5], %[ftmp5], %[ftmp9] \n\t" + "psrah %[ftmp3], %[ftmp3], %[ftmp9] \n\t" + "psubh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "psubh %[ftmp2], %[ftmp2], %[ftmp3] \n\t" + "mov.d %[ftmp3], %[ftmp13] \n\t" + "psrah %[ftmp0], %[ftmp13], %[ftmp8] \n\t" + "psrah %[ftmp7], %[ftmp6], %[ftmp8] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp6] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" + MMI_LDC1(%[ftmp6], $29, 0x08) + "dmtc1 %[tmp6], %[ftmp3] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp3] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp6] \n\t" + "psubh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp7] \n\t" + "paddh %[ftmp0], %[ftmp0], %[ftmp0] \n\t" + "paddh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" + "psubh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp7] \n\t" + "paddh %[ftmp1], %[ftmp1], %[ftmp6] \n\t" + "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp6] \n\t" + "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" + "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" + "paddh %[ftmp3], %[ftmp3], %[ftmp3] \n\t" + MMI_SDC1(%[ftmp6], $29, 0x08) + "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" + MMI_SDC1(%[ftmp7], $29, 0x18) + "dmfc1 %[tmp2], %[ftmp0] \n\t" + "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" + MMI_ULWC1(%[ftmp6], %[addr0], 0x00) + MMI_LWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + "psrah %[ftmp2], %[ftmp2], %[ftmp10] \n\t" + "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" + "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" + "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + MMI_SWC1(%[ftmp6], %[addr0], 0x00) + MMI_SWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + MMI_ULWC1(%[ftmp6], %[addr0], 0x00) + MMI_LWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" + "psrah %[ftmp4], %[ftmp4], %[ftmp10] \n\t" + "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" + "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + MMI_SWC1(%[ftmp6], %[addr0], 0x00) + MMI_SWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + MMI_LDC1(%[ftmp2], $29, 0x08) + MMI_LDC1(%[ftmp5], $29, 0x18) + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + "dmtc1 %[tmp2], %[ftmp1] \n\t" + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + MMI_ULWC1(%[ftmp6], %[addr0], 0x00) + MMI_LWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + "psrah %[ftmp3], %[ftmp3], %[ftmp10] \n\t" + "psrah %[ftmp2], %[ftmp2], %[ftmp10] \n\t" + "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" + "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + MMI_SWC1(%[ftmp6], %[addr0], 0x00) + MMI_SWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" + MMI_ULWC1(%[ftmp6], %[addr0], 0x00) + MMI_LWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + "psrah %[ftmp5], %[ftmp5], %[ftmp10] \n\t" + "psrah %[ftmp1], %[ftmp1], %[ftmp10] \n\t" + "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "punpcklbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" + "paddh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" + "packushb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" + "packushb %[ftmp7], %[ftmp7], %[ftmp0] \n\t" + MMI_SWC1(%[ftmp6], %[addr0], 0x00) + MMI_SWXC1(%[ftmp7], %[addr0], %[stride], 0x00) + PTR_ADDIU "$29, $29, 0x20 \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), @@ -648,9 +628,11 @@ void ff_h264_idct8_add_8_mmi(uint8_t *dst, int16_t *block, int stride) [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), [tmp2]"=&r"(tmp[2]), [tmp3]"=&r"(tmp[3]), [tmp4]"=&r"(tmp[4]), [tmp5]"=&r"(tmp[5]), - [tmp6]"=&r"(tmp[6]), [tmp7]"=&r"(tmp[7]), - [addr0]"=&r"(addr[0]), - [low32]"=&r"(low32) + [tmp6]"=&r"(tmp[6]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT + [addr0]"=&r"(addr[0]) : [dst]"r"(dst), [block]"r"(block), [stride]"r"((mips_reg)stride) : "$29","memory" @@ -663,7 +645,7 @@ void ff_h264_idct_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) { int dc = (block[0] + 32) >> 6; double ftmp[6]; - uint64_t low32; + DECLARE_VAR_LOW32; block[0] = 0; @@ -671,14 +653,10 @@ void ff_h264_idct_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "mtc1 %[dc], %[ftmp5] \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "pshufh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[dst0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[dst1]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[dst2]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x00(%[dst3]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_ULWC1(%[ftmp1], %[dst0], 0x00) + MMI_ULWC1(%[ftmp2], %[dst1], 0x00) + MMI_ULWC1(%[ftmp3], %[dst2], 0x00) + MMI_ULWC1(%[ftmp4], %[dst3], 0x00) "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" @@ -691,18 +669,15 @@ void ff_h264_idct_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "packushb %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "packushb %[ftmp3], %[ftmp3], %[ftmp0] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp0] \n\t" - "gsswlc1 %[ftmp1], 0x03(%[dst0]) \n\t" - "gsswrc1 %[ftmp1], 0x00(%[dst0]) \n\t" - "gsswlc1 %[ftmp2], 0x03(%[dst1]) \n\t" - "gsswrc1 %[ftmp2], 0x00(%[dst1]) \n\t" - "gsswlc1 %[ftmp3], 0x03(%[dst2]) \n\t" - "gsswrc1 %[ftmp3], 0x00(%[dst2]) \n\t" - "gsswlc1 %[ftmp4], 0x03(%[dst3]) \n\t" - "gsswrc1 %[ftmp4], 0x00(%[dst3]) \n\t" + MMI_SWC1(%[ftmp1], %[dst0], 0x00) + MMI_SWC1(%[ftmp2], %[dst1], 0x00) + MMI_SWC1(%[ftmp3], %[dst2], 0x00) + MMI_SWC1(%[ftmp4], %[dst3], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), - [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), - [low32]"=&r"(low32) + [ftmp4]"=&f"(ftmp[4]), + RESTRICT_ASM_LOW32 + [ftmp5]"=&f"(ftmp[5]) : [dst0]"r"(dst), [dst1]"r"(dst+stride), [dst2]"r"(dst+2*stride), [dst3]"r"(dst+3*stride), [dc]"r"(dc) @@ -714,6 +689,7 @@ void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) { int dc = (block[0] + 32) >> 6; double ftmp[10]; + DECLARE_VAR_ALL64; block[0] = 0; @@ -721,10 +697,10 @@ void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "mtc1 %[dc], %[ftmp5] \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "pshufh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" - "ldc1 %[ftmp1], 0x00(%[dst0]) \n\t" - "ldc1 %[ftmp2], 0x00(%[dst1]) \n\t" - "ldc1 %[ftmp3], 0x00(%[dst2]) \n\t" - "ldc1 %[ftmp4], 0x00(%[dst3]) \n\t" + MMI_LDC1(%[ftmp1], %[dst0], 0x00) + MMI_LDC1(%[ftmp2], %[dst1], 0x00) + MMI_LDC1(%[ftmp3], %[dst2], 0x00) + MMI_LDC1(%[ftmp4], %[dst3], 0x00) "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t" @@ -745,15 +721,15 @@ void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "packushb %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "packushb %[ftmp3], %[ftmp3], %[ftmp8] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp9] \n\t" - "sdc1 %[ftmp1], 0x00(%[dst0]) \n\t" - "sdc1 %[ftmp2], 0x00(%[dst1]) \n\t" - "sdc1 %[ftmp3], 0x00(%[dst2]) \n\t" - "sdc1 %[ftmp4], 0x00(%[dst3]) \n\t" - - "ldc1 %[ftmp1], 0x00(%[dst4]) \n\t" - "ldc1 %[ftmp2], 0x00(%[dst5]) \n\t" - "ldc1 %[ftmp3], 0x00(%[dst6]) \n\t" - "ldc1 %[ftmp4], 0x00(%[dst7]) \n\t" + MMI_SDC1(%[ftmp1], %[dst0], 0x00) + MMI_SDC1(%[ftmp2], %[dst1], 0x00) + MMI_SDC1(%[ftmp3], %[dst2], 0x00) + MMI_SDC1(%[ftmp4], %[dst3], 0x00) + + MMI_LDC1(%[ftmp1], %[dst4], 0x00) + MMI_LDC1(%[ftmp2], %[dst5], 0x00) + MMI_LDC1(%[ftmp3], %[dst6], 0x00) + MMI_LDC1(%[ftmp4], %[dst7], 0x00) "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t" @@ -774,15 +750,17 @@ void ff_h264_idct8_dc_add_8_mmi(uint8_t *dst, int16_t *block, int stride) "packushb %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "packushb %[ftmp3], %[ftmp3], %[ftmp8] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp9] \n\t" - "sdc1 %[ftmp1], 0x00(%[dst4]) \n\t" - "sdc1 %[ftmp2], 0x00(%[dst5]) \n\t" - "sdc1 %[ftmp3], 0x00(%[dst6]) \n\t" - "sdc1 %[ftmp4], 0x00(%[dst7]) \n\t" + MMI_SDC1(%[ftmp1], %[dst4], 0x00) + MMI_SDC1(%[ftmp2], %[dst5], 0x00) + MMI_SDC1(%[ftmp3], %[dst6], 0x00) + MMI_SDC1(%[ftmp4], %[dst7], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), - [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]) + [ftmp8]"=&f"(ftmp[8]), + RESTRICT_ASM_ALL64 + [ftmp9]"=&f"(ftmp[9]) : [dst0]"r"(dst), [dst1]"r"(dst+stride), [dst2]"r"(dst+2*stride), [dst3]"r"(dst+3*stride), [dst4]"r"(dst+4*stride), [dst5]"r"(dst+5*stride), @@ -888,17 +866,18 @@ void ff_h264_luma_dc_dequant_idct_8_mmi(int16_t *output, int16_t *input, { double ftmp[10]; uint64_t tmp[2]; + DECLARE_VAR_ALL64; __asm__ volatile ( ".set noreorder \n\t" "dli %[tmp0], 0x08 \n\t" - "ldc1 %[ftmp3], 0x18(%[input]) \n\t" + MMI_LDC1(%[ftmp3], %[input], 0x18) "mtc1 %[tmp0], %[ftmp8] \n\t" - "ldc1 %[ftmp2], 0x10(%[input]) \n\t" + MMI_LDC1(%[ftmp2], %[input], 0x10) "dli %[tmp0], 0x20 \n\t" - "ldc1 %[ftmp1], 0x08(%[input]) \n\t" + MMI_LDC1(%[ftmp1], %[input], 0x08) "mtc1 %[tmp0], %[ftmp9] \n\t" - "ldc1 %[ftmp0], 0x00(%[input]) \n\t" + MMI_LDC1(%[ftmp0], %[input], 0x00) "mov.d %[ftmp4], %[ftmp3] \n\t" "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" "psubh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" @@ -1009,7 +988,10 @@ void ff_h264_luma_dc_dequant_idct_8_mmi(int16_t *output, int16_t *input, "sh %[input], 0x1e0(%[output]) \n\t" "1: \n\t" "ori %[tmp0], $0, 0x1f \n\t" +#if HAVE_LOONGSON3 "clz %[tmp1], %[qmul] \n\t" +#elif HAVE_LOONGSON2 +#endif "ori %[input], $0, 0x07 \n\t" "dsubu %[tmp1], %[tmp0], %[tmp1] \n\t" "ori %[tmp0], $0, 0x80 \n\t" @@ -1098,6 +1080,7 @@ void ff_h264_luma_dc_dequant_idct_8_mmi(int16_t *output, int16_t *input, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), + RESTRICT_ASM_ALL64 [output]"+&r"(output), [input]"+&r"(input), [qmul]"+&r"(qmul) : [ff_pw_1]"f"(ff_pw_1) @@ -1157,6 +1140,7 @@ void ff_h264_weight_pixels16_8_mmi(uint8_t *block, ptrdiff_t stride, int height, { int y; double ftmp[8]; + DECLARE_VAR_ALL64; offset <<= log2_denom; @@ -1166,8 +1150,8 @@ void ff_h264_weight_pixels16_8_mmi(uint8_t *block, ptrdiff_t stride, int height, for (y=0; y> 2; const int l7 = (src[-1+6*stride] + 2*src[-1+7*stride] + src[-1+7*stride] + 2) >> 2; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; + __asm__ volatile ( - "gsldlc1 %[ftmp4], 0x07(%[srcA]) \n\t" - "gsldrc1 %[ftmp4], 0x00(%[srcA]) \n\t" - "gsldlc1 %[ftmp5], 0x07(%[src0]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[src0]) \n\t" - "gsldlc1 %[ftmp6], 0x07(%[src1]) \n\t" - "gsldrc1 %[ftmp6], 0x00(%[src1]) \n\t" + MMI_ULDC1(%[ftmp4], %[srcA], 0x00) + MMI_ULDC1(%[ftmp5], %[src0], 0x00) + MMI_ULDC1(%[ftmp6], %[src1], 0x00) "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x03 \n\t" "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t" @@ -309,7 +309,9 @@ void ff_pred8x8l_dc_8_mmi(uint8_t *src, int has_topleft, int has_topright, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [ftmp11]"=&f"(ftmp[11]), [ftmp12]"=&f"(ftmp[12]), [ftmp13]"=&f"(ftmp[13]), - [tmp0]"=&r"(tmp[0]), [dc2]"=r"(dc2) + [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + [dc2]"=r"(dc2) : [srcA]"r"((mips_reg)(src-stride-1)), [src0]"r"((mips_reg)(src-stride)), [src1]"r"((mips_reg)(src-stride+1)), @@ -323,20 +325,22 @@ void ff_pred8x8l_dc_8_mmi(uint8_t *src, int has_topleft, int has_topright, __asm__ volatile ( "dli %[tmp0], 0x02 \n\t" "punpcklwd %[ftmp0], %[dc], %[dc] \n\t" + "1: \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" - "gssdxc1 %[ftmp0], 0x00(%[src], %[stride]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) + MMI_SDXC1(%[ftmp0], %[src], %[stride], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" - "gssdxc1 %[ftmp0], 0x00(%[src], %[stride]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) + MMI_SDXC1(%[ftmp0], %[src], %[stride], 0x00) + "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" "bnez %[tmp0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [src]"+&r"(src) : [dc]"f"(dc), [stride]"r"((mips_reg)stride) : "memory" @@ -348,15 +352,13 @@ void ff_pred8x8l_vertical_8_mmi(uint8_t *src, int has_topleft, { double ftmp[12]; mips_reg tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "gsldlc1 %[ftmp3], 0x07(%[srcA]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[srcA]) \n\t" - "gsldlc1 %[ftmp4], 0x07(%[src0]) \n\t" - "gsldrc1 %[ftmp4], 0x00(%[src0]) \n\t" - "gsldlc1 %[ftmp5], 0x07(%[src1]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[src1]) \n\t" + MMI_LDC1(%[ftmp3], %[srcA], 0x00) + MMI_LDC1(%[ftmp4], %[src0], 0x00) + MMI_LDC1(%[ftmp5], %[src1], 0x00) "punpcklbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t" "punpckhbh %[ftmp7], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t" @@ -385,7 +387,7 @@ void ff_pred8x8l_vertical_8_mmi(uint8_t *src, int has_topleft, "psrah %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp7], %[ftmp7], %[ftmp1] \n\t" "packushb %[ftmp4], %[ftmp6], %[ftmp7] \n\t" - "sdc1 %[ftmp4], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp4], %[src], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), @@ -393,6 +395,7 @@ void ff_pred8x8l_vertical_8_mmi(uint8_t *src, int has_topleft, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [ftmp11]"=&f"(ftmp[11]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [src]"=r"(src) : [srcA]"r"((mips_reg)(src-stride-1)), [src0]"r"((mips_reg)(src-stride)), @@ -403,22 +406,21 @@ void ff_pred8x8l_vertical_8_mmi(uint8_t *src, int has_topleft, __asm__ volatile ( "dli %[tmp0], 0x02 \n\t" + "1: \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) + "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" "bnez %[tmp0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [src]"+&r"(src) : [stride]"r"((mips_reg)stride) : "memory" @@ -433,19 +435,21 @@ void ff_pred4x4_dc_8_mmi(uint8_t *src, const uint8_t *topright, + src[-1+2*stride] + src[-1+3*stride] + 4) >>3; uint64_t tmp[2]; mips_reg addr[1]; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[tmp0], %[dc], $0 \n\t" "dmul %[tmp1], %[tmp0], %[ff_pb_1] \n\t" "xor %[addr0], %[addr0], %[addr0] \n\t" - "gsswx %[tmp1], 0x00(%[src], %[addr0]) \n\t" + MMI_SWX(%[tmp1], %[src], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "gsswx %[tmp1], 0x00(%[src], %[addr0]) \n\t" + MMI_SWX(%[tmp1], %[src], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "gsswx %[tmp1], 0x00(%[src], %[addr0]) \n\t" + MMI_SWX(%[tmp1], %[src], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "gsswx %[tmp1], 0x00(%[src], %[addr0]) \n\t" + MMI_SWX(%[tmp1], %[src], %[addr0], 0x00) : [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]) : [src]"r"((mips_reg)src), [stride]"r"((mips_reg)stride), [dc]"r"(dc), [ff_pb_1]"r"(ff_pb_1) @@ -518,13 +522,13 @@ void ff_pred8x8_top_dc_8_mmi(uint8_t *src, ptrdiff_t stride) double ftmp[4]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "dli %[tmp0], 0x02 \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" PTR_SUBU "%[addr0], %[src], %[stride] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" + MMI_LDC1(%[ftmp1], %[addr0], 0x00) "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "biadd %[ftmp2], %[ftmp2] \n\t" @@ -539,32 +543,25 @@ void ff_pred8x8_top_dc_8_mmi(uint8_t *src, ptrdiff_t stride) "psrlh %[ftmp2], %[ftmp2], %[ftmp1] \n\t" "psrlh %[ftmp3], %[ftmp3], %[ftmp1] \n\t" "packushb %[ftmp1], %[ftmp2], %[ftmp3] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp1], %[src], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [src]"+&r"(src) : [stride]"r"((mips_reg)stride) @@ -651,21 +648,21 @@ void ff_pred8x8_dc_8_mmi(uint8_t *src, ptrdiff_t stride) "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "packushb %[ftmp2], %[ftmp3], %[ftmp4] \n\t" PTR_ADDU "%[addr0], $0, %[src] \n\t" - "sdc1 %[ftmp1], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp1], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp1], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp1], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp2], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp2], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp2], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp2], %[addr0], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), @@ -682,28 +679,27 @@ void ff_pred8x16_vertical_8_mmi(uint8_t *src, ptrdiff_t stride) { double ftmp[1]; uint64_t tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( - "gsldlc1 %[ftmp0], 0x07(%[srcA]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[srcA]) \n\t" + MMI_LDC1(%[ftmp0], %[srcA], 0x00) "dli %[tmp0], 0x04 \n\t" + "1: \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) PTR_ADDU "%[src], %[src], %[stride] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_SDC1(%[ftmp0], %[src], 0x00) + "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[stride] \n\t" "bnez %[tmp0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [src]"+&r"(src) : [stride]"r"((mips_reg)stride), [srcA]"r"((mips_reg)(src-stride)) : "memory" @@ -747,17 +743,16 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, const int svq3, const int rv40) { double ftmp[11]; - uint64_t tmp[7]; + uint64_t tmp[6]; mips_reg addr[1]; + DECLARE_VAR_ALL64; __asm__ volatile( PTR_SUBU "%[addr0], %[src], %[stride] \n\t" - "dli %[tmp2], 0x20 \n\t" - "dmtc1 %[tmp2], %[ftmp4] \n\t" - "gsldlc1 %[ftmp0], 0x06(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp0], -0x01(%[addr0]) \n\t" - "gsldrc1 %[ftmp2], 0x08(%[addr0]) \n\t" + "dli %[tmp0], 0x20 \n\t" + "dmtc1 %[tmp0], %[ftmp4] \n\t" + MMI_ULDC1(%[ftmp0], %[addr0], -0x01) + MMI_ULDC1(%[ftmp2], %[addr0], 0x08) "dsrl %[ftmp1], %[ftmp0], %[ftmp4] \n\t" "dsrl %[ftmp3], %[ftmp2], %[ftmp4] \n\t" "xor %[ftmp4], %[ftmp4], %[ftmp4] \n\t" @@ -772,29 +767,29 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, "paddsh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "paddsh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "paddsh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "dli %[tmp2], 0x0e \n\t" - "dmtc1 %[tmp2], %[ftmp4] \n\t" + "dli %[tmp0], 0x0e \n\t" + "dmtc1 %[tmp0], %[ftmp4] \n\t" "pshufh %[ftmp1], %[ftmp0], %[ftmp4] \n\t" "paddsh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "dli %[tmp2], 0x01 \n\t" - "dmtc1 %[tmp2], %[ftmp4] \n\t" + "dli %[tmp0], 0x01 \n\t" + "dmtc1 %[tmp0], %[ftmp4] \n\t" "pshufh %[ftmp1], %[ftmp0], %[ftmp4] \n\t" "paddsh %[ftmp5], %[ftmp0], %[ftmp1] \n\t" PTR_ADDIU "%[addr0], %[src], -0x01 \n\t" PTR_SUBU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp2], 0x00(%[addr0]) \n\t" - "lbu %[tmp6], 0x10(%[addr0]) \n\t" + "lbu %[tmp5], 0x10(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp3], 0x00(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp4], 0x00(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "lbu %[tmp5], 0x00(%[addr0]) \n\t" + "lbu %[tmp0], 0x00(%[addr0]) \n\t" "dsll %[tmp3], %[tmp3], 0x10 \n\t" "dsll %[tmp4], %[tmp4], 0x20 \n\t" - "dsll %[tmp5], %[tmp5], 0x30 \n\t" - "or %[tmp4], %[tmp4], %[tmp5] \n\t" + "dsll %[tmp0], %[tmp0], 0x30 \n\t" + "or %[tmp4], %[tmp4], %[tmp0] \n\t" "or %[tmp2], %[tmp2], %[tmp3] \n\t" "or %[tmp2], %[tmp2], %[tmp4] \n\t" "dmtc1 %[tmp2], %[ftmp0] \n\t" @@ -806,11 +801,11 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp4], 0x00(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "lbu %[tmp5], 0x00(%[addr0]) \n\t" + "lbu %[tmp0], 0x00(%[addr0]) \n\t" "dsll %[tmp3], %[tmp3], 0x10 \n\t" "dsll %[tmp4], %[tmp4], 0x20 \n\t" - "dsll %[tmp5], %[tmp5], 0x30 \n\t" - "or %[tmp4], %[tmp4], %[tmp5] \n\t" + "dsll %[tmp0], %[tmp0], 0x30 \n\t" + "or %[tmp4], %[tmp4], %[tmp0] \n\t" "or %[tmp2], %[tmp2], %[tmp3] \n\t" "or %[tmp2], %[tmp2], %[tmp4] \n\t" "dmtc1 %[tmp2], %[ftmp1] \n\t" @@ -823,11 +818,11 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp4], 0x00(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "lbu %[tmp5], 0x00(%[addr0]) \n\t" + "lbu %[tmp0], 0x00(%[addr0]) \n\t" "dsll %[tmp3], %[tmp3], 0x10 \n\t" "dsll %[tmp4], %[tmp4], 0x20 \n\t" - "dsll %[tmp5], %[tmp5], 0x30 \n\t" - "or %[tmp4], %[tmp4], %[tmp5] \n\t" + "dsll %[tmp0], %[tmp0], 0x30 \n\t" + "or %[tmp4], %[tmp4], %[tmp0] \n\t" "or %[tmp2], %[tmp2], %[tmp3] \n\t" "or %[tmp2], %[tmp2], %[tmp4] \n\t" "dmtc1 %[tmp2], %[ftmp2] \n\t" @@ -839,15 +834,15 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" "lbu %[tmp4], 0x00(%[addr0]) \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" - "lbu %[tmp5], 0x00(%[addr0]) \n\t" - "daddu %[tmp6], %[tmp6], %[tmp5] \n\t" - "daddiu %[tmp6], %[tmp6], 0x01 \n\t" - "dsll %[tmp6], %[tmp6], 0x04 \n\t" + "lbu %[tmp0], 0x00(%[addr0]) \n\t" + "daddu %[tmp5], %[tmp5], %[tmp0] \n\t" + "daddiu %[tmp5], %[tmp5], 0x01 \n\t" + "dsll %[tmp5], %[tmp5], 0x04 \n\t" "dsll %[tmp3], %[tmp3], 0x10 \n\t" "dsll %[tmp4], %[tmp4], 0x20 \n\t" - "dsll %[tmp5], %[tmp5], 0x30 \n\t" - "or %[tmp4], %[tmp4], %[tmp5] \n\t" + "dsll %[tmp0], %[tmp0], 0x30 \n\t" + "or %[tmp4], %[tmp4], %[tmp0] \n\t" "or %[tmp2], %[tmp2], %[tmp3] \n\t" "or %[tmp2], %[tmp2], %[tmp4] \n\t" "dmtc1 %[tmp2], %[ftmp3] \n\t" @@ -859,13 +854,13 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, "paddsh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "paddsh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "paddsh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "dli %[tmp2], 0x0e \n\t" - "dmtc1 %[tmp2], %[ftmp4] \n\t" + "dli %[tmp0], 0x0e \n\t" + "dmtc1 %[tmp0], %[ftmp4] \n\t" "pshufh %[ftmp1], %[ftmp0], %[ftmp4] \n\t" "paddsh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "dli %[tmp2], 0x01 \n\t" - "dmtc1 %[tmp2], %[ftmp4] \n\t" + "dli %[tmp0], 0x01 \n\t" + "dmtc1 %[tmp0], %[ftmp4] \n\t" "pshufh %[ftmp1], %[ftmp0], %[ftmp4] \n\t" "paddsh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" @@ -914,17 +909,17 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, "daddu %[tmp3], %[tmp0], %[tmp1] \n\t" "dli %[tmp2], 0x07 \n\t" "dmul %[tmp3], %[tmp3], %[tmp2] \n\t" - "dsubu %[tmp6], %[tmp6], %[tmp3] \n\t" + "dsubu %[tmp5], %[tmp5], %[tmp3] \n\t" "xor %[ftmp4], %[ftmp4], %[ftmp4] \n\t" "dmtc1 %[tmp0], %[ftmp0] \n\t" "pshufh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "dmtc1 %[tmp1], %[ftmp5] \n\t" "pshufh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" - "dmtc1 %[tmp6], %[ftmp6] \n\t" + "dmtc1 %[tmp5], %[ftmp6] \n\t" "pshufh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "dli %[tmp2], 0x05 \n\t" - "dmtc1 %[tmp2], %[ftmp7] \n\t" + "dli %[tmp0], 0x05 \n\t" + "dmtc1 %[tmp0], %[ftmp7] \n\t" "pmullh %[ftmp1], %[ff_pw_0to3], %[ftmp0] \n\t" "dmtc1 %[ff_pw_4to7], %[ftmp2] \n\t" "pmullh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" @@ -941,16 +936,14 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, "paddsh %[ftmp9], %[ftmp2], %[ftmp6] \n\t" "psrah %[ftmp9], %[ftmp9], %[ftmp7] \n\t" "packushb %[ftmp0], %[ftmp8], %[ftmp9] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[addr0]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[addr0]) \n\t" + MMI_SDC1(%[ftmp0], %[addr0], 0x00) "paddsh %[ftmp8], %[ftmp3], %[ftmp6] \n\t" "psrah %[ftmp8], %[ftmp8], %[ftmp7] \n\t" "paddsh %[ftmp9], %[ftmp4], %[ftmp6] \n\t" "psrah %[ftmp9], %[ftmp9], %[ftmp7] \n\t" "packushb %[ftmp0], %[ftmp8], %[ftmp9] \n\t" - "gssdlc1 %[ftmp0], 0x0f(%[addr0]) \n\t" - "gssdrc1 %[ftmp0], 0x08(%[addr0]) \n\t" + MMI_SDC1(%[ftmp0], %[addr0], 0x08) "paddsh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" PTR_ADDU "%[addr0], %[addr0], %[stride] \n\t" @@ -964,7 +957,7 @@ static inline void pred16x16_plane_compat_mmi(uint8_t *src, int stride, [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), [tmp2]"=&r"(tmp[2]), [tmp3]"=&r"(tmp[3]), [tmp4]"=&r"(tmp[4]), [tmp5]"=&r"(tmp[5]), - [tmp6]"=&r"(tmp[6]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]) : [src]"r"(src), [stride]"r"((mips_reg)stride), [svq3]"r"(svq3), [rv40]"r"(rv40), diff --git a/libavcodec/mips/h264qpel_mmi.c b/libavcodec/mips/h264qpel_mmi.c index b4e83e4..13fbebf 100644 --- a/libavcodec/mips/h264qpel_mmi.c +++ b/libavcodec/mips/h264qpel_mmi.c @@ -24,28 +24,26 @@ #include "h264dsp_mips.h" #include "hpeldsp_mips.h" #include "libavcodec/bit_depth_template.c" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" static inline void copy_block4_mmi(uint8_t *dst, const uint8_t *src, int dstStride, int srcStride, int h) { double ftmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; __asm__ volatile ( "1: \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "gsswlc1 %[ftmp0], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_ULWC1(%[ftmp0], %[src], 0x00) + MMI_SWC1(%[ftmp0], %[dst], 0x00) "addi %[h], %[h], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [dst]"+&r"(dst), [src]"+&r"(src), - [h]"+&r"(h), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride) : "memory" @@ -56,18 +54,18 @@ static inline void copy_block8_mmi(uint8_t *dst, const uint8_t *src, int dstStride, int srcStride, int h) { double ftmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src]) \n\t" - "gssdlc1 %[ftmp0], 0x07(%[dst]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_ULDC1(%[ftmp0], %[src], 0x00) + MMI_SDC1(%[ftmp0], %[dst], 0x00) "addi %[h], %[h], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride), @@ -81,15 +79,14 @@ static inline void copy_block16_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[1]; uint64_t tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_ULDC1(%[ftmp0], %[src], 0x00) "ldl %[tmp0], 0x0f(%[src]) \n\t" "ldr %[tmp0], 0x08(%[src]) \n\t" - "gssdlc1 %[ftmp0], 0x07(%[dst]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) "sdl %[tmp0], 0x0f(%[dst]) \n\t" "sdr %[tmp0], 0x08(%[dst]) \n\t" "addi %[h], %[h], -0x01 \n\t" @@ -98,6 +95,7 @@ static inline void copy_block16_mmi(uint8_t *dst, const uint8_t *src, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src), [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride), @@ -113,24 +111,19 @@ static void put_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[10]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x04 \n\t" "1: \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], -0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" - "uld %[low32], 0x03(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp1], %[src], -0x02) + MMI_ULWC1(%[ftmp2], %[src], -0x01) + MMI_ULWC1(%[ftmp3], %[src], 0x00) + MMI_ULWC1(%[ftmp4], %[src], 0x01) + MMI_ULWC1(%[ftmp5], %[src], 0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x03) + "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" @@ -147,8 +140,7 @@ static void put_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddsh %[ftmp9], %[ftmp9], %[ff_pw_16] \n\t" "psrah %[ftmp9], %[ftmp9], %[ff_pw_5] \n\t" "packushb %[ftmp9], %[ftmp9], %[ftmp0] \n\t" - "gsswlc1 %[ftmp9], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp9], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp9], %[dst], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" @@ -159,8 +151,8 @@ static void put_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), - [dst]"+&r"(dst), [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [dst]"+&r"(dst), [src]"+&r"(src) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), [ff_pw_20]"f"(ff_pw_20), [ff_pw_5]"f"(ff_pw_5), @@ -174,23 +166,18 @@ static void put_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[11]; uint64_t tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x08 \n\t" "1: \n\t" - "gsldlc1 %[ftmp1], 0x05(%[src]) \n\t" - "gsldrc1 %[ftmp1], -0x02(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x06(%[src]) \n\t" - "gsldrc1 %[ftmp2], -0x01(%[src]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp4], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[src]) \n\t" - "gsldlc1 %[ftmp5], 0x09(%[src]) \n\t" - "gsldrc1 %[ftmp5], 0x02(%[src]) \n\t" - "gsldlc1 %[ftmp6], 0x0a(%[src]) \n\t" - "gsldrc1 %[ftmp6], 0x03(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], -0x02) + MMI_ULDC1(%[ftmp2], %[src], -0x01) + MMI_ULDC1(%[ftmp3], %[src], 0x00) + MMI_ULDC1(%[ftmp4], %[src], 0x01) + MMI_ULDC1(%[ftmp5], %[src], 0x02) + MMI_ULDC1(%[ftmp6], %[src], 0x03) "punpcklbh %[ftmp7], %[ftmp3], %[ftmp0] \n\t" "punpckhbh %[ftmp8], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp9], %[ftmp4], %[ftmp0] \n\t" @@ -222,8 +209,7 @@ static void put_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, "psrah %[ftmp3], %[ftmp3], %[ff_pw_5] \n\t" "psrah %[ftmp4], %[ftmp4], %[ff_pw_5] \n\t" "packushb %[ftmp9], %[ftmp3], %[ftmp4] \n\t" - "gssdlc1 %[ftmp9], 0x07(%[dst]) \n\t" - "gssdrc1 %[ftmp9], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp9], %[dst], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" @@ -235,6 +221,7 @@ static void put_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), @@ -260,24 +247,18 @@ static void avg_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[11]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x04 \n\t" "1: \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], -0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" - "uld %[low32], 0x03(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp1], %[src], -0x02) + MMI_ULWC1(%[ftmp2], %[src], -0x01) + MMI_ULWC1(%[ftmp3], %[src], 0x00) + MMI_ULWC1(%[ftmp4], %[src], 0x01) + MMI_ULWC1(%[ftmp5], %[src], 0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x03) "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" @@ -294,10 +275,9 @@ static void avg_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddsh %[ftmp9], %[ftmp9], %[ff_pw_16] \n\t" "psrah %[ftmp9], %[ftmp9], %[ff_pw_5] \n\t" "packushb %[ftmp9], %[ftmp9], %[ftmp0] \n\t" - "lwc1 %[ftmp10], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp10], %[dst], 0x00) "pavgb %[ftmp9], %[ftmp9], %[ftmp10] \n\t" - "gsswlc1 %[ftmp9], 0x03(%[dst]) \n\t" - "gsswrc1 %[ftmp9], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp9], %[dst], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" @@ -309,8 +289,8 @@ static void avg_h264_qpel4_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [tmp0]"=&r"(tmp[0]), - [dst]"+&r"(dst), [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [dst]"+&r"(dst), [src]"+&r"(src) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), [ff_pw_20]"f"(ff_pw_20), [ff_pw_5]"f"(ff_pw_5), @@ -324,23 +304,18 @@ static void avg_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[11]; uint64_t tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x08 \n\t" "1: \n\t" - "gsldlc1 %[ftmp1], 0x05(%[src]) \n\t" - "gsldrc1 %[ftmp1], -0x02(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x06(%[src]) \n\t" - "gsldrc1 %[ftmp2], -0x01(%[src]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp4], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[src]) \n\t" - "gsldlc1 %[ftmp5], 0x09(%[src]) \n\t" - "gsldrc1 %[ftmp5], 0x02(%[src]) \n\t" - "gsldlc1 %[ftmp6], 0x0a(%[src]) \n\t" - "gsldrc1 %[ftmp6], 0x03(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], -0x02) + MMI_ULDC1(%[ftmp2], %[src], -0x01) + MMI_ULDC1(%[ftmp3], %[src], 0x00) + MMI_ULDC1(%[ftmp4], %[src], 0x01) + MMI_ULDC1(%[ftmp5], %[src], 0x02) + MMI_ULDC1(%[ftmp6], %[src], 0x03) "punpcklbh %[ftmp7], %[ftmp3], %[ftmp0] \n\t" "punpckhbh %[ftmp8], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp9], %[ftmp4], %[ftmp0] \n\t" @@ -372,9 +347,9 @@ static void avg_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, "psrah %[ftmp3], %[ftmp3], %[ff_pw_5] \n\t" "psrah %[ftmp4], %[ftmp4], %[ff_pw_5] \n\t" "packushb %[ftmp9], %[ftmp3], %[ftmp4] \n\t" - "ldc1 %[ftmp10], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp10], %[dst], 0x00) "pavgb %[ftmp9], %[ftmp9], %[ftmp10] \n\t" - "sdc1 %[ftmp9], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp9], %[dst], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" @@ -386,6 +361,7 @@ static void avg_h264_qpel8_h_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), @@ -411,7 +387,7 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[12]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2 * srcStride; @@ -420,31 +396,25 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, ".set noreorder \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x02 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "mtc1 %[tmp0], %[ftmp10] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "dli %[tmp0], 0x05 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "mtc1 %[tmp0], %[ftmp11] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp0] \n\t" "punpcklbh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_LWC1(%[ftmp6], %[src], 0x00) "paddh %[ftmp7], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp7], %[ftmp7], %[ftmp10] \n\t" "psubh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" @@ -457,10 +427,9 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" "psrah %[ftmp7], %[ftmp7], %[ftmp11] \n\t" "packushb %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "swc1 %[ftmp7], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp7], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp7], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp7], %[ftmp7], %[ftmp10] \n\t" "psubh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" @@ -473,10 +442,9 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp7], %[ftmp7], %[ftmp2] \n\t" "psrah %[ftmp7], %[ftmp7], %[ftmp11] \n\t" "packushb %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "swc1 %[ftmp7], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp7], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp7], %[ftmp5], %[ftmp6] \n\t" "psllh %[ftmp7], %[ftmp7], %[ftmp10] \n\t" "psubh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" @@ -489,10 +457,9 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp7], %[ftmp7], %[ftmp3] \n\t" "psrah %[ftmp7], %[ftmp7], %[ftmp11] \n\t" "packushb %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "swc1 %[ftmp7], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp7], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) "paddh %[ftmp7], %[ftmp6], %[ftmp1] \n\t" "psllh %[ftmp7], %[ftmp7], %[ftmp10] \n\t" "psubh %[ftmp7], %[ftmp7], %[ftmp5] \n\t" @@ -505,7 +472,7 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp7], %[ftmp7], %[ftmp4] \n\t" "psrah %[ftmp7], %[ftmp7], %[ftmp11] \n\t" "packushb %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "swc1 %[ftmp7], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp7], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" ".set pop \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), @@ -515,8 +482,8 @@ static void put_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [ftmp11]"=&f"(ftmp[11]), [tmp0]"=&r"(tmp[0]), - [dst]"+&r"(dst), [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [dst]"+&r"(dst), [src]"+&r"(src) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -531,7 +498,7 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, int h = 8; double ftmp[10]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2 * srcStride; @@ -540,29 +507,23 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, ".set push \n\t" ".set noreorder \n\t" "dli %[tmp0], 0x02 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "mtc1 %[tmp0], %[ftmp8] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "dli %[tmp0], 0x05 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "mtc1 %[tmp0], %[ftmp9] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" @@ -578,10 +539,9 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -594,11 +554,10 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" @@ -610,12 +569,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" @@ -626,12 +584,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "punpcklbh %[ftmp3] , %[ftmp3], %[ftmp7] \n\t" @@ -642,12 +599,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" @@ -658,12 +614,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "punpcklbh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" @@ -674,12 +629,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -690,12 +644,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "bne %[h], 0x10, 2f \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" @@ -707,12 +660,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" @@ -723,12 +675,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" @@ -739,12 +690,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" @@ -755,12 +705,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "punpcklbh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" @@ -771,12 +720,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -787,12 +735,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" @@ -803,12 +750,11 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" @@ -819,7 +765,7 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "2: \n\t" ".set pop \n\t" @@ -829,9 +775,9 @@ static void put_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [src]"+&r"(src), [dst]"+&r"(dst), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -869,23 +815,23 @@ static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" "mtc1 %[tmp0], %[ftmp9] \n\t" "dli %[tmp0], 0x05 \n\t" - "lwc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "mtc1 %[tmp0], %[ftmp8] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "lwc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "lwc1 %[ftmp2], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "lwc1 %[ftmp3], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "lwc1 %[ftmp4], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "lwc1 %[ftmp5], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" @@ -898,11 +844,11 @@ static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp0], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "lwc1 %[ftmp0], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" @@ -915,11 +861,11 @@ static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp1], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "lwc1 %[ftmp1], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" @@ -932,11 +878,11 @@ static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "lwc1 %[ftmp2], 0x00(%[src]) \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" @@ -949,9 +895,9 @@ static void avg_h264_qpel4_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp3], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp3], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" ".set pop \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), @@ -975,7 +921,7 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, int h = 8; double ftmp[10]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2 * srcStride; @@ -987,29 +933,23 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" "mtc1 %[tmp0], %[ftmp9] \n\t" "dli %[tmp0], 0x05 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "mtc1 %[tmp0], %[ftmp8] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" @@ -1022,12 +962,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp0], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" @@ -1040,12 +979,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp1], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" @@ -1058,12 +996,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" @@ -1076,12 +1013,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp3], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp3], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" @@ -1094,12 +1030,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp4], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp4], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" @@ -1112,12 +1047,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp5], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp5], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" @@ -1130,12 +1064,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp0], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" @@ -1148,13 +1081,12 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp1], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) "bne %[h], 0x10, 2f \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" @@ -1167,12 +1099,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" @@ -1185,12 +1116,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp3], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp3], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_LWC1(%[ftmp3], %[src], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" @@ -1203,12 +1133,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp4], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp4], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_LWC1(%[ftmp4], %[src], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" @@ -1221,12 +1150,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp5], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp5], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_LWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" @@ -1239,12 +1167,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp0], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_LWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" @@ -1257,12 +1184,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp1], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_LWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" @@ -1275,12 +1201,11 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp2], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp2], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_LWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "psubh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" @@ -1293,9 +1218,9 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" "psrah %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp6] \n\t" - "lwc1 %[ftmp3], 0x00(%[dst]) \n\t" + MMI_LWC1(%[ftmp3], %[dst], 0x00) "pavgb %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "swc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_SWC1(%[ftmp6], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "2: \n\t" ".set pop \n\t" @@ -1305,9 +1230,9 @@ static void avg_h264_qpel8_v_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 [src]"+&r"(src), [dst]"+&r"(dst), - [h]"+&r"(h), - [low32]"=&r"(low32) + [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride), [srcStride]"r"((mips_reg)srcStride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -1339,7 +1264,7 @@ static void put_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, int16_t *tmp = _tmp; double ftmp[10]; uint64_t tmp0; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2*srcStride; @@ -1347,18 +1272,12 @@ static void put_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x09 \n\t" "1: \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], -0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" - "uld %[low32], 0x03(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp1], %[src], -0x02) + MMI_ULWC1(%[ftmp2], %[src], -0x01) + MMI_ULWC1(%[ftmp3], %[src], 0x00) + MMI_ULWC1(%[ftmp4], %[src], 0x01) + MMI_ULWC1(%[ftmp5], %[src], 0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x03) "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" @@ -1372,7 +1291,7 @@ static void put_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, "pmullh %[ftmp8], %[ftmp8], %[ff_pw_5] \n\t" "psubsh %[ftmp7], %[ftmp7], %[ftmp8] \n\t" "paddsh %[ftmp9], %[ftmp7], %[ftmp9] \n\t" - "sdc1 %[ftmp9], 0x00(%[tmp]) \n\t" + MMI_SDC1(%[ftmp9], %[tmp], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[tmp], %[tmp], %[tmpStride] \n\t" @@ -1383,8 +1302,8 @@ static void put_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp0), - [tmp]"+&r"(tmp), [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [tmp]"+&r"(tmp), [src]"+&r"(src) : [tmpStride]"r"(8), [srcStride]"r"((mips_reg)srcStride), [ff_pw_20]"f"(ff_pw_20), [ff_pw_5]"f"(ff_pw_5) @@ -1418,37 +1337,31 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, int w = (size + 8) >> 2; double ftmp[11]; uint64_t tmp0; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2 * srcStride + 2; while (w--) { __asm__ volatile ( "dli %[tmp0], 0x02 \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_ULWC1(%[ftmp0], %[src], 0x00) "mtc1 %[tmp0], %[ftmp10] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp2], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp3], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_ULWC1(%[ftmp4], %[src], 0x00) PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" "punpcklbh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_ULWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp0], %[ftmp0], %[ff_pw_16] \n\t" @@ -1459,9 +1372,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp0], %[ftmp0], %[ftmp5] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "sdc1 %[ftmp6], 0x00(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x00) + MMI_ULWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_16] \n\t" @@ -1472,9 +1384,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "sdc1 %[ftmp6], 0x30(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x30) + MMI_ULWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp2], %[ftmp2], %[ff_pw_16] \n\t" @@ -1485,9 +1396,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp2], %[ftmp2], %[ftmp1] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "sdc1 %[ftmp6], 0x60(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x60) + MMI_ULWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp3], %[ftmp3], %[ff_pw_16] \n\t" @@ -1498,9 +1408,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "sdc1 %[ftmp6], 0x90(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x90) + MMI_ULWC1(%[ftmp3], %[src], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp4], %[ftmp4], %[ff_pw_16] \n\t" @@ -1511,9 +1420,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "sdc1 %[ftmp6], 0xc0(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0xc0) + MMI_ULWC1(%[ftmp4], %[src], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp5], %[ftmp5], %[ff_pw_16] \n\t" @@ -1524,9 +1432,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "sdc1 %[ftmp6], 0xf0(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0xf0) + MMI_ULWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp0], %[ftmp0], %[ff_pw_16] \n\t" @@ -1537,9 +1444,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp0], %[ftmp0], %[ftmp5] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "sdc1 %[ftmp6], 0x120(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x120) + MMI_ULWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_16] \n\t" @@ -1550,11 +1456,10 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "sdc1 %[ftmp6], 0x150(%[tmp]) \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x150) "bne %[size], 0x10, 2f \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp2], %[ftmp2], %[ff_pw_16] \n\t" @@ -1565,9 +1470,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp2], %[ftmp2], %[ftmp1] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "sdc1 %[ftmp6], 0x180(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x180) + MMI_ULWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp3], %[ftmp3], %[ff_pw_16] \n\t" @@ -1578,9 +1482,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "sdc1 %[ftmp6], 0x1b0(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x1b0) + MMI_ULWC1(%[ftmp3], %[src], 0x00) "paddh %[ftmp6], %[ftmp0], %[ftmp1] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp4], %[ftmp4], %[ff_pw_16] \n\t" @@ -1591,9 +1494,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp4], %[ftmp4], %[ftmp3] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp4] \n\t" - "sdc1 %[ftmp6], 0x1e0(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x1e0) + MMI_ULWC1(%[ftmp4], %[src], 0x00) "paddh %[ftmp6], %[ftmp1], %[ftmp2] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp5], %[ftmp5], %[ff_pw_16] \n\t" @@ -1604,9 +1506,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp5], %[ftmp5], %[ftmp4] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp5] \n\t" - "sdc1 %[ftmp6], 0x210(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x210) + MMI_ULWC1(%[ftmp5], %[src], 0x00) "paddh %[ftmp6], %[ftmp2], %[ftmp3] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp0], %[ftmp0], %[ff_pw_16] \n\t" @@ -1617,9 +1518,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp0], %[ftmp0], %[ftmp5] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" - "sdc1 %[ftmp6], 0x240(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x240) + MMI_ULWC1(%[ftmp0], %[src], 0x00) "paddh %[ftmp6], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp1], %[ftmp1], %[ff_pw_16] \n\t" @@ -1630,9 +1530,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" - "sdc1 %[ftmp6], 0x270(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x270) + MMI_ULWC1(%[ftmp1], %[src], 0x00) "paddh %[ftmp6], %[ftmp4], %[ftmp5] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp2], %[ftmp2], %[ff_pw_16] \n\t" @@ -1643,9 +1542,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp2], %[ftmp2], %[ftmp1] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp2] \n\t" - "sdc1 %[ftmp6], 0x2a0(%[tmp]) \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x2a0) + MMI_ULWC1(%[ftmp2], %[src], 0x00) "paddh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" "psllh %[ftmp6], %[ftmp6], %[ftmp10] \n\t" "paddh %[ftmp3], %[ftmp3], %[ff_pw_16] \n\t" @@ -1656,7 +1554,7 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, "paddh %[ftmp3], %[ftmp3], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" "paddh %[ftmp6], %[ftmp6], %[ftmp3] \n\t" - "sdc1 %[ftmp6], 0x2d0(%[tmp]) \n\t" + MMI_SDC1(%[ftmp6], %[tmp], 0x2d0) "2: \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), @@ -1665,8 +1563,8 @@ static void put_h264_qpel8or16_hv1_lowpass_mmi(int16_t *tmp, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [tmp0]"=&r"(tmp0), - [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [src]"+&r"(src) : [tmp]"r"(tmp), [size]"r"(size), [srcStride]"r"((mips_reg)srcStride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -1684,6 +1582,7 @@ static void put_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, int w = size >> 4; double ftmp[10]; uint64_t tmp0; + DECLARE_VAR_ALL64; do { int h = size; @@ -1694,27 +1593,20 @@ static void put_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, "dli %[tmp0], 0x06 \n\t" "mtc1 %[tmp0], %[ftmp9] \n\t" "1: \n\t" - "ldc1 %[ftmp0], 0x00(%[tmp]) \n\t" - "ldc1 %[ftmp3], 0x08(%[tmp]) \n\t" - "ldc1 %[ftmp6], 0x10(%[tmp]) \n\t" - "gsldlc1 %[ftmp1], 0x09(%[tmp]) \n\t" - "gsldrc1 %[ftmp1], 0x02(%[tmp]) \n\t" - "gsldlc1 %[ftmp4], 0x11(%[tmp]) \n\t" - "gsldrc1 %[ftmp4], 0x0a(%[tmp]) \n\t" - "gsldlc1 %[ftmp5], 0x19(%[tmp]) \n\t" - "gsldrc1 %[ftmp5], 0x12(%[tmp]) \n\t" + MMI_LDC1(%[ftmp0], %[tmp], 0x00) + MMI_LDC1(%[ftmp3], %[tmp], 0x08) + MMI_LDC1(%[ftmp6], %[tmp], 0x10) + MMI_ULDC1(%[ftmp1], %[tmp], 0x02) + MMI_ULDC1(%[ftmp4], %[tmp], 0x0a) + MMI_ULDC1(%[ftmp5], %[tmp], 0x12) "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" "paddh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" - "gsldlc1 %[ftmp2], 0x0b(%[tmp]) \n\t" - "gsldrc1 %[ftmp2], 0x04(%[tmp]) \n\t" - "gsldlc1 %[ftmp6], 0x0d(%[tmp]) \n\t" - "gsldrc1 %[ftmp6], 0x06(%[tmp]) \n\t" - "gsldlc1 %[ftmp5], 0x13(%[tmp]) \n\t" - "gsldrc1 %[ftmp5], 0x0c(%[tmp]) \n\t" - "gsldlc1 %[ftmp7], 0x15(%[tmp]) \n\t" - "gsldrc1 %[ftmp7], 0x0e(%[tmp]) \n\t" + MMI_ULDC1(%[ftmp2], %[tmp], 0x04) + MMI_ULDC1(%[ftmp6], %[tmp], 0x06) + MMI_ULDC1(%[ftmp5], %[tmp], 0x0c) + MMI_ULDC1(%[ftmp7], %[tmp], 0x0e) "paddh %[ftmp2], %[ftmp2], %[ftmp6] \n\t" "paddh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" "psubh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" @@ -1733,8 +1625,7 @@ static void put_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, "psrah %[ftmp3], %[ftmp3], %[ftmp9] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp3] \n\t" "addi %[h], %[h], -0x01 \n\t" - "gssdlc1 %[ftmp0], 0x07(%[dst]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) PTR_ADDIU "%[tmp], %[tmp], 0x30 \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" "bnez %[h], 1b \n\t" @@ -1744,6 +1635,7 @@ static void put_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp0), + RESTRICT_ASM_ALL64 [tmp]"+&r"(tmp), [dst]"+&r"(dst), [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride) @@ -1785,7 +1677,8 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, int h = 8; double ftmp[9]; uint64_t tmp[1]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ALL64; __asm__ volatile ( "dli %[tmp0], 0x02 \n\t" @@ -1794,10 +1687,8 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "mtc1 %[tmp0], %[ftmp8] \n\t" "1: \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp3], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp3], 0x01(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp3], %[src], 0x01) "punpckhbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp4], %[ftmp3], %[ftmp0] \n\t" @@ -1806,10 +1697,8 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "psllh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "psllh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" - "gsldlc1 %[ftmp3], 0x06(%[src]) \n\t" - "gsldrc1 %[ftmp3], -0x01(%[src]) \n\t" - "gsldlc1 %[ftmp5], 0x09(%[src]) \n\t" - "gsldrc1 %[ftmp5], 0x02(%[src]) \n\t" + MMI_ULDC1(%[ftmp3], %[src], -0x01) + MMI_ULDC1(%[ftmp5], %[src], 0x02) "punpckhbh %[ftmp4], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" "punpckhbh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" @@ -1820,10 +1709,8 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "psubh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pmullh %[ftmp2], %[ftmp2], %[ff_pw_5] \n\t" "pmullh %[ftmp1], %[ftmp1], %[ff_pw_5] \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x07(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp3], %[src], -0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x07) "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "paddh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" @@ -1834,13 +1721,12 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp2], %[ftmp2], %[ftmp5] \n\t" "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" "psrah %[ftmp2], %[ftmp2], %[ftmp8] \n\t" - "gsldlc1 %[ftmp5], 0x07(%[src2]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[src2]) \n\t" + MMI_LDC1(%[ftmp5], %[src2], 0x00) "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" PTR_ADDU "%[src], %[src], %[dstStride] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" PTR_ADDU "%[h], %[h], -0x01 \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" PTR_ADDU "%[src2], %[src2], %[src2Stride] \n\t" "bgtz %[h], 1b \n\t" @@ -1850,9 +1736,10 @@ static void put_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ALL64 [src]"+&r"(src), [dst]"+&r"(dst), - [src2]"+&r"(src2), [h]"+&r"(h), - [low32]"=&r"(low32) + [src2]"+&r"(src2), [h]"+&r"(h) : [src2Stride]"r"((mips_reg)src2Stride), [dstStride]"r"((mips_reg)dstStride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -1865,35 +1752,35 @@ static void put_pixels8_l2_shift5_mmi(uint8_t *dst, int16_t *src16, { double ftmp[7]; uint64_t tmp0; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; do { __asm__ volatile ( "dli %[tmp0], 0x05 \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src16]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src16]) \n\t" + MMI_ULDC1(%[ftmp0], %[src16], 0x00) "mtc1 %[tmp0], %[ftmp6] \n\t" - "gsldlc1 %[ftmp1], 0x0f(%[src16]) \n\t" - "gsldrc1 %[ftmp1], 0x08(%[src16]) \n\t" - "gsldlc1 %[ftmp2], 0x37(%[src16]) \n\t" - "gsldrc1 %[ftmp2], 0x30(%[src16]) \n\t" - "gsldlc1 %[ftmp3], 0x3f(%[src16]) \n\t" - "gsldrc1 %[ftmp3], 0x38(%[src16]) \n\t" + MMI_ULDC1(%[ftmp1], %[src16], 0x08) + MMI_ULDC1(%[ftmp2], %[src16], 0x30) + MMI_ULDC1(%[ftmp3], %[src16], 0x38) "psrah %[ftmp0], %[ftmp0], %[ftmp6] \n\t" "psrah %[ftmp1], %[ftmp1], %[ftmp6] \n\t" "psrah %[ftmp2], %[ftmp2], %[ftmp6] \n\t" "psrah %[ftmp3], %[ftmp3], %[ftmp6] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp1] \n\t" "packushb %[ftmp2], %[ftmp2], %[ftmp3] \n\t" - "ldc1 %[ftmp5], 0x00(%[src8]) \n\t" - "gsldxc1 %[ftmp4], 0x00(%[src8], %[src8Stride]) \n\t" + MMI_LDC1(%[ftmp5], %[src8], 0x00) + MMI_LDXC1(%[ftmp4], %[src8], %[src8Stride], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp5] \n\t" "pavgb %[ftmp2], %[ftmp2], %[ftmp4] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp2], 0x00(%[dst], %[dstStride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp2], %[dst], %[dstStride], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [tmp0]"=&r"(tmp0) : [src8]"r"(src8), [src16]"r"(src16), [dst]"r"(dst), @@ -1941,7 +1828,7 @@ static void avg_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, int16_t *tmp = _tmp; double ftmp[10]; uint64_t tmp0; - uint64_t low32; + DECLARE_VAR_LOW32; src -= 2*srcStride; @@ -1949,18 +1836,12 @@ static void avg_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "dli %[tmp0], 0x09 \n\t" "1: \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], -0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[src]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" - "uld %[low32], 0x01(%[src]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" - "uld %[low32], 0x03(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp1], %[src], -0x02) + MMI_ULWC1(%[ftmp2], %[src], -0x01) + MMI_ULWC1(%[ftmp3], %[src], 0x00) + MMI_ULWC1(%[ftmp4], %[src], 0x01) + MMI_ULWC1(%[ftmp5], %[src], 0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x03) "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp3], %[ftmp3], %[ftmp0] \n\t" @@ -1974,7 +1855,7 @@ static void avg_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, "pmullh %[ftmp8], %[ftmp8], %[ff_pw_5] \n\t" "psubsh %[ftmp7], %[ftmp7], %[ftmp8] \n\t" "paddsh %[ftmp9], %[ftmp7], %[ftmp9] \n\t" - "sdc1 %[ftmp9], 0x00(%[tmp]) \n\t" + MMI_SDC1(%[ftmp9], %[tmp], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[src], %[src], %[srcStride] \n\t" PTR_ADDU "%[tmp], %[tmp], %[tmpStride] \n\t" @@ -1985,8 +1866,8 @@ static void avg_h264_qpel4_hv_lowpass_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp0), - [tmp]"+&r"(tmp), [src]"+&r"(src), - [low32]"=&r"(low32) + RESTRICT_ASM_LOW32 + [tmp]"+&r"(tmp), [src]"+&r"(src) : [tmpStride]"r"(8), [srcStride]"r"((mips_reg)srcStride), [ff_pw_20]"f"(ff_pw_20), [ff_pw_5]"f"(ff_pw_5) @@ -2020,6 +1901,7 @@ static void avg_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, int w = size >> 4; double ftmp[11]; uint64_t tmp0; + DECLARE_VAR_ALL64; do { int h = size; @@ -2029,27 +1911,20 @@ static void avg_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, "dli %[tmp0], 0x06 \n\t" "mtc1 %[tmp0], %[ftmp10] \n\t" "1: \n\t" - "ldc1 %[ftmp0], 0x00(%[tmp]) \n\t" - "ldc1 %[ftmp3], 0x08(%[tmp]) \n\t" - "gsldlc1 %[ftmp1], 0x09(%[tmp]) \n\t" - "gsldrc1 %[ftmp1], 0x02(%[tmp]) \n\t" - "gsldlc1 %[ftmp4], 0x11(%[tmp]) \n\t" - "gsldrc1 %[ftmp4], 0x0a(%[tmp]) \n\t" - "ldc1 %[ftmp7], 0x10(%[tmp]) \n\t" - "gsldlc1 %[ftmp8], 0x19(%[tmp]) \n\t" - "gsldrc1 %[ftmp8], 0x12(%[tmp]) \n\t" + MMI_LDC1(%[ftmp0], %[tmp], 0x00) + MMI_LDC1(%[ftmp3], %[tmp], 0x08) + MMI_ULDC1(%[ftmp1], %[tmp], 0x02) + MMI_ULDC1(%[ftmp4], %[tmp], 0x0a) + MMI_LDC1(%[ftmp7], %[tmp], 0x10) + MMI_ULDC1(%[ftmp8], %[tmp], 0x12) "paddh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "paddh %[ftmp3], %[ftmp3], %[ftmp8] \n\t" "paddh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" - "gsldlc1 %[ftmp2], 0x0b(%[tmp]) \n\t" - "gsldrc1 %[ftmp2], 0x04(%[tmp]) \n\t" - "gsldlc1 %[ftmp5], 0x13(%[tmp]) \n\t" - "gsldrc1 %[ftmp5], 0x0c(%[tmp]) \n\t" - "gsldlc1 %[ftmp7], 0x0d(%[tmp]) \n\t" - "gsldrc1 %[ftmp7], 0x06(%[tmp]) \n\t" - "gsldlc1 %[ftmp8], 0x15(%[tmp]) \n\t" - "gsldrc1 %[ftmp8], 0x0e(%[tmp]) \n\t" + MMI_ULDC1(%[ftmp2], %[tmp], 0x04) + MMI_ULDC1(%[ftmp5], %[tmp], 0x0c) + MMI_ULDC1(%[ftmp7], %[tmp], 0x06) + MMI_ULDC1(%[ftmp8], %[tmp], 0x0e) "paddh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" "paddh %[ftmp5], %[ftmp5], %[ftmp8] \n\t" "psubh %[ftmp0], %[ftmp0], %[ftmp1] \n\t" @@ -2067,9 +1942,9 @@ static void avg_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, "psrah %[ftmp0], %[ftmp0], %[ftmp10] \n\t" "psrah %[ftmp3], %[ftmp3], %[ftmp10] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp3] \n\t" - "ldc1 %[ftmp6], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp6], %[dst], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp6] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) "addi %[h], %[h], -0x01 \n\t" PTR_ADDI "%[tmp], %[tmp], 0x30 \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" @@ -2081,6 +1956,7 @@ static void avg_h264_qpel8or16_hv2_lowpass_mmi(uint8_t *dst, [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [ftmp10]"=&f"(ftmp[10]), [tmp0]"=&r"(tmp0), + RESTRICT_ASM_ALL64 [tmp]"+&r"(tmp), [dst]"+&r"(dst), [h]"+&r"(h) : [dstStride]"r"((mips_reg)dstStride) @@ -2121,7 +1997,8 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, { double ftmp[10]; uint64_t tmp[2]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ALL64; __asm__ volatile ( "dli %[tmp1], 0x02 \n\t" @@ -2131,10 +2008,8 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "mtc1 %[tmp1], %[ftmp8] \n\t" "1: \n\t" - "gsldlc1 %[ftmp1], 0x07(%[src]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[src]) \n\t" - "gsldlc1 %[ftmp2], 0x08(%[src]) \n\t" - "gsldrc1 %[ftmp2], 0x01(%[src]) \n\t" + MMI_ULDC1(%[ftmp1], %[src], 0x00) + MMI_ULDC1(%[ftmp2], %[src], 0x01) "punpckhbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" "punpcklbh %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpckhbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" @@ -2143,10 +2018,8 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" "psllh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" "psllh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" - "gsldlc1 %[ftmp2], 0x06(%[src]) \n\t" - "gsldrc1 %[ftmp2], -0x01(%[src]) \n\t" - "gsldlc1 %[ftmp5], 0x09(%[src]) \n\t" - "gsldrc1 %[ftmp5], 0x02(%[src]) \n\t" + MMI_ULDC1(%[ftmp2], %[src], -0x01) + MMI_ULDC1(%[ftmp5], %[src], 0x02) "punpckhbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpckhbh %[ftmp6], %[ftmp5], %[ftmp0] \n\t" @@ -2157,10 +2030,8 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "psubh %[ftmp3], %[ftmp3], %[ftmp6] \n\t" "pmullh %[ftmp1], %[ftmp1], %[ff_pw_5] \n\t" "pmullh %[ftmp3], %[ftmp3], %[ff_pw_5] \n\t" - "uld %[low32], -0x02(%[src]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x07(%[src]) \n\t" - "mtc1 %[low32], %[ftmp6] \n\t" + MMI_ULWC1(%[ftmp2], %[src], -0x02) + MMI_ULWC1(%[ftmp6], %[src], 0x07) "punpcklbh %[ftmp2], %[ftmp2], %[ftmp0] \n\t" "punpcklbh %[ftmp6], %[ftmp6], %[ftmp0] \n\t" "paddh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" @@ -2171,14 +2042,13 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" "psrah %[ftmp1], %[ftmp1], %[ftmp8] \n\t" "psrah %[ftmp3], %[ftmp3], %[ftmp8] \n\t" - "gsldlc1 %[ftmp5], 0x07(%[src2]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[src2]) \n\t" + MMI_LDC1(%[ftmp5], %[src2], 0x00) "packushb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "ldc1 %[ftmp9], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp9], %[dst], 0x00) "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp9] \n\t" PTR_ADDU "%[src], %[src], %[dstStride] \n\t" - "sdc1 %[ftmp1], 0x00(%[dst]) \n\t" + MMI_SDC1(%[ftmp1], %[dst], 0x00) "daddi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDU "%[dst], %[dst], %[dstStride] \n\t" PTR_ADDU "%[src2], %[src2], %[src2Stride] \n\t" @@ -2189,9 +2059,10 @@ static void avg_h264_qpel8_h_lowpass_l2_mmi(uint8_t *dst, const uint8_t *src, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ALL64 [dst]"+&r"(dst), [src]"+&r"(src), - [src2]"+&r"(src2), - [low32]"=&r"(low32) + [src2]"+&r"(src2) : [dstStride]"r"((mips_reg)dstStride), [src2Stride]"r"((mips_reg)src2Stride), [ff_pw_5]"f"(ff_pw_5), [ff_pw_16]"f"(ff_pw_16) @@ -2220,39 +2091,39 @@ static void avg_pixels8_l2_shift5_mmi(uint8_t *dst, int16_t *src16, { double ftmp[8]; uint64_t tmp0; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; do { __asm__ volatile ( "dli %[tmp0], 0x05 \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src16]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src16]) \n\t" + MMI_ULDC1(%[ftmp0], %[src16], 0x00) "mtc1 %[tmp0], %[ftmp6] \n\t" - "gsldlc1 %[ftmp1], 0x0f(%[src16]) \n\t" - "gsldrc1 %[ftmp1], 0x08(%[src16]) \n\t" - "gsldlc1 %[ftmp2], 0x37(%[src16]) \n\t" - "gsldrc1 %[ftmp2], 0x30(%[src16]) \n\t" - "gsldlc1 %[ftmp3], 0x3f(%[src16]) \n\t" - "gsldrc1 %[ftmp3], 0x38(%[src16]) \n\t" + MMI_ULDC1(%[ftmp1], %[src16], 0x08) + MMI_ULDC1(%[ftmp2], %[src16], 0x30) + MMI_ULDC1(%[ftmp3], %[src16], 0x38) "psrah %[ftmp0], %[ftmp0], %[ftmp6] \n\t" "psrah %[ftmp1], %[ftmp1], %[ftmp6] \n\t" "psrah %[ftmp2], %[ftmp2], %[ftmp6] \n\t" "psrah %[ftmp3], %[ftmp3], %[ftmp6] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "ldc1 %[ftmp4], 0x00(%[src8]) \n\t" - "gsldxc1 %[ftmp5], 0x00(%[src8], %[src8Stride]) \n\t" + MMI_LDC1(%[ftmp4], %[src8], 0x00) + MMI_LDXC1(%[ftmp5], %[src8], %[src8Stride], 0x00) "packushb %[ftmp2], %[ftmp2], %[ftmp3] \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "pavgb %[ftmp2], %[ftmp2], %[ftmp5] \n\t" - "ldc1 %[ftmp7], 0x00(%[dst]) \n\t" + MMI_LDC1(%[ftmp7], %[dst], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gsldxc1 %[ftmp7], 0x00(%[dst], %[dstStride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_LDXC1(%[ftmp7], %[dst], %[dstStride], 0x00) "pavgb %[ftmp2], %[ftmp2], %[ftmp7] \n\t" - "gssdxc1 %[ftmp2], 0x00(%[dst], %[dstStride]) \n\t" + MMI_SDXC1(%[ftmp2], %[dst], %[dstStride], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [tmp0]"=&r"(tmp0) : [src8]"r"(src8), [src16]"r"(src16), [dst]"r"(dst), diff --git a/libavcodec/mips/hpeldsp_mmi.c b/libavcodec/mips/hpeldsp_mmi.c index 4c46f00..2dbef22 100644 --- a/libavcodec/mips/hpeldsp_mmi.c +++ b/libavcodec/mips/hpeldsp_mmi.c @@ -23,7 +23,7 @@ #include "hpeldsp_mips.h" #include "libavcodec/bit_depth_template.c" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" #include "constants.h" void ff_put_pixels4_8_mmi(uint8_t *block, const uint8_t *pixels, @@ -31,36 +31,34 @@ void ff_put_pixels4_8_mmi(uint8_t *block, const uint8_t *pixels, { double ftmp[2]; mips_reg addr[2]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr1], %[line_size], %[line_size] \n\t" "1: \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "uld %[low32], 0x00(%[pixels]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "swc1 %[ftmp0], 0x00(%[block]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_ULWC1(%[ftmp0], %[pixels], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_SWC1(%[ftmp0], %[block], 0x00) + MMI_SWXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "uld %[low32], 0x00(%[pixels]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "swc1 %[ftmp0], 0x00(%[block]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_ULWC1(%[ftmp0], %[pixels], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_SWC1(%[ftmp0], %[block], 0x00) + MMI_SWXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" PTR_ADDI "%[h], %[h], -0x04 \n\t" "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), - [low32]"=&r"(low32), [block]"+&r"(block), [pixels]"+&r"(pixels), [h]"+&r"(h) : [line_size]"r"((mips_reg)line_size) @@ -72,35 +70,36 @@ void ff_put_pixels8_8_mmi(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h) { double ftmp[2]; - mips_reg addr[2]; + mips_reg addr[3]; + DECLARE_VAR_ALL64; __asm__ volatile ( PTR_ADDU "%[addr1], %[line_size], %[line_size] \n\t" "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_SDC1(%[ftmp0], %[block], 0x00) + PTR_ADDU "%[addr2], %[block], %[line_size] \n\t" + MMI_SDC1(%[ftmp1], %[addr2], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_SDC1(%[ftmp0], %[block], 0x00) + PTR_ADDU "%[addr2], %[block], %[line_size] \n\t" + MMI_SDC1(%[ftmp1], %[addr2], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" PTR_ADDI "%[h], %[h], -0x04 \n\t" "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), + [addr2]"=&r"(addr[2]), [block]"+&r"(block), [pixels]"+&r"(pixels), [h]"+&r"(h) : [line_size]"r"((mips_reg)line_size) @@ -113,39 +112,33 @@ void ff_put_pixels16_8_mmi(uint8_t *block, const uint8_t *pixels, { double ftmp[4]; mips_reg addr[2]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr1], %[line_size], %[line_size] \n\t" "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp2], 0x0f(%[pixels]) \n\t" - "gsldrc1 %[ftmp2], 0x08(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp3], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp3], 0x08(%[addr0]) \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" - "sdc1 %[ftmp2], 0x08(%[block]) \n\t" - "gssdxc1 %[ftmp3], 0x08(%[block], %[line_size]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) + MMI_ULDC1(%[ftmp2], %[pixels], 0x08) + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp3], %[addr0], 0x08) + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) + MMI_SDC1(%[ftmp2], %[block], 0x08) + MMI_SDXC1(%[ftmp3], %[block], %[line_size], 0x08) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp2], 0x0f(%[pixels]) \n\t" - "gsldrc1 %[ftmp2], 0x08(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp3], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp3], 0x08(%[addr0]) \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" - "sdc1 %[ftmp2], 0x08(%[block]) \n\t" - "gssdxc1 %[ftmp3], 0x08(%[block], %[line_size]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) + MMI_ULDC1(%[ftmp2], %[pixels], 0x08) + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp3], %[addr0], 0x08) + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) + MMI_SDC1(%[ftmp2], %[block], 0x08) + MMI_SDXC1(%[ftmp3], %[block], %[line_size], 0x08) PTR_ADDU "%[pixels], %[pixels], %[addr1] \n\t" PTR_ADDU "%[block], %[block], %[addr1] \n\t" @@ -153,6 +146,8 @@ void ff_put_pixels16_8_mmi(uint8_t *block, const uint8_t *pixels, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [block]"+&r"(block), [pixels]"+&r"(pixels), [h]"+&r"(h) @@ -166,42 +161,35 @@ void ff_avg_pixels4_8_mmi(uint8_t *block, const uint8_t *pixels, { double ftmp[4]; mips_reg addr[3]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[line_size], %[line_size] \n\t" "1: \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "uld %[low32], 0x00(%[pixels]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp0], %[pixels], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "uld %[low32], 0x00(%[block]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp2], %[block], 0x00) + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "swc1 %[ftmp0], 0x00(%[block]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_SWC1(%[ftmp0], %[block], 0x00) + MMI_SWXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "uld %[low32], 0x00(%[pixels]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" + MMI_ULWC1(%[ftmp0], %[pixels], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "uld %[low32], 0x00(%[block]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp2], %[block], 0x00) + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "swc1 %[ftmp0], 0x00(%[block]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_SWC1(%[ftmp0], %[block], 0x00) + MMI_SWXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" @@ -209,9 +197,10 @@ void ff_avg_pixels4_8_mmi(uint8_t *block, const uint8_t *pixels, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), - [low32]"=&r"(low32), [block]"+&r"(block), [pixels]"+&r"(pixels), [h]"+&r"(h) : [line_size]"r"((mips_reg)line_size) @@ -224,41 +213,35 @@ void ff_avg_pixels8_8_mmi(uint8_t *block, const uint8_t *pixels, { double ftmp[4]; mips_reg addr[3]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[line_size], %[line_size] \n\t" "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[block]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[block]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp2], %[block], 0x00) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[block]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[block]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp2], %[block], 0x00) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" @@ -266,6 +249,8 @@ void ff_avg_pixels8_8_mmi(uint8_t *block, const uint8_t *pixels, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [block]"+&r"(block), [pixels]"+&r"(pixels), @@ -280,65 +265,51 @@ void ff_avg_pixels16_8_mmi(uint8_t *block, const uint8_t *pixels, { double ftmp[8]; mips_reg addr[3]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[line_size], %[line_size] \n\t" "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp4], 0x0f(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp4], %[pixels], 0x08) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "gsldrc1 %[ftmp4], 0x08(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp5], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp5], 0x08(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[block]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[block]) \n\t" - "gsldlc1 %[ftmp6], 0x0f(%[block]) \n\t" - "gsldrc1 %[ftmp6], 0x08(%[block]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp7], 0x0f(%[addr1]) \n\t" - "gsldrc1 %[ftmp7], 0x08(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp5], %[addr0], 0x08) + MMI_ULDC1(%[ftmp2], %[block], 0x00) + MMI_ULDC1(%[ftmp6], %[block], 0x08) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) + MMI_ULDC1(%[ftmp7], %[addr1], 0x08) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp4], %[ftmp4], %[ftmp6] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pavgb %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" - "sdc1 %[ftmp4], 0x08(%[block]) \n\t" - "gssdxc1 %[ftmp5], 0x08(%[block], %[line_size]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) + MMI_SDC1(%[ftmp4], %[block], 0x08) + MMI_SDXC1(%[ftmp5], %[block], %[line_size], 0x08) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "gsldlc1 %[ftmp4], 0x0f(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp4], %[pixels], 0x08) PTR_ADDU "%[addr1], %[block], %[line_size] \n\t" - "gsldrc1 %[ftmp4], 0x08(%[pixels]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp5], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp5], 0x08(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[block]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[block]) \n\t" - "gsldlc1 %[ftmp6], 0x0f(%[block]) \n\t" - "gsldrc1 %[ftmp6], 0x08(%[block]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp7], 0x0f(%[addr1]) \n\t" - "gsldrc1 %[ftmp7], 0x08(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp5], %[addr0], 0x08) + MMI_ULDC1(%[ftmp2], %[block], 0x00) + MMI_ULDC1(%[ftmp6], %[block], 0x08) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) + MMI_ULDC1(%[ftmp7], %[addr1], 0x08) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp4], %[ftmp4], %[ftmp6] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pavgb %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[block], %[line_size]) \n\t" - "sdc1 %[ftmp4], 0x08(%[block]) \n\t" - "gssdxc1 %[ftmp5], 0x08(%[block], %[line_size]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDXC1(%[ftmp1], %[block], %[line_size], 0x00) + MMI_SDC1(%[ftmp4], %[block], 0x08) + MMI_SDXC1(%[ftmp5], %[block], %[line_size], 0x08) PTR_ADDU "%[pixels], %[pixels], %[addr2] \n\t" PTR_ADDU "%[block], %[block], %[addr2] \n\t" @@ -348,6 +319,8 @@ void ff_avg_pixels16_8_mmi(uint8_t *block, const uint8_t *pixels, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [block]"+&r"(block), [pixels]"+&r"(pixels), @@ -363,7 +336,8 @@ inline void ff_put_pixels4_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[4]; mips_reg addr[5]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" @@ -371,38 +345,30 @@ inline void ff_put_pixels4_l2_8_mmi(uint8_t *dst, const uint8_t *src1, PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" "1: \n\t" PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "uld %[low32], 0x00(%[src1]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[src2]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp0], %[src1], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_ULWC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "swc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SWC1(%[ftmp0], %[dst], 0x00) + MMI_SWXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "uld %[low32], 0x00(%[src1]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[src2]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp0], %[src1], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_ULWC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "swc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SWC1(%[ftmp0], %[dst], 0x00) + MMI_SWXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -410,10 +376,11 @@ inline void ff_put_pixels4_l2_8_mmi(uint8_t *dst, const uint8_t *src1, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), - [low32]"=&r"(low32), [dst]"+&r"(dst), [src1]"+&r"(src1), [src2]"+&r"(src2), [h]"+&r"(h) : [dst_stride]"r"((mips_reg)dst_stride), @@ -429,45 +396,40 @@ inline void ff_put_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[4]; mips_reg addr[5]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" PTR_ADDU "%[addr3], %[src_stride2], %[src_stride2] \n\t" PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -475,6 +437,8 @@ inline void ff_put_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, "bnez %[h], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), @@ -493,69 +457,56 @@ inline void ff_put_pixels16_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[8]; mips_reg addr[5]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" PTR_ADDU "%[addr3], %[src_stride2], %[src_stride2] \n\t" PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp4], 0x0f(%[src1]) \n\t" - "gsldrc1 %[ftmp4], 0x08(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp5], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp5], 0x08(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp4], %[src1], 0x08) + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp5], %[addr0], 0x08) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp6], 0x0f(%[src2]) \n\t" - "gsldrc1 %[ftmp6], 0x08(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp6], %[src2], 0x08) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp7], 0x0f(%[addr1]) \n\t" - "gsldrc1 %[ftmp7], 0x08(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp7], %[addr1], 0x08) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp4], %[ftmp4], %[ftmp6] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pavgb %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" - "sdc1 %[ftmp4], 0x08(%[dst]) \n\t" - "gssdxc1 %[ftmp5], 0x08(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) + MMI_SDC1(%[ftmp4], %[dst], 0x08) + MMI_SDXC1(%[ftmp5], %[dst], %[dst_stride], 0x08) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp4], 0x0f(%[src1]) \n\t" - "gsldrc1 %[ftmp4], 0x08(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp5], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp5], 0x08(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp4], %[src1], 0x08) + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp5], %[addr0], 0x08) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp6], 0x0f(%[src2]) \n\t" - "gsldrc1 %[ftmp6], 0x08(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp6], %[src2], 0x08) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp7], 0x0f(%[addr1]) \n\t" - "gsldrc1 %[ftmp7], 0x08(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp7], %[addr1], 0x08) "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp4], %[ftmp4], %[ftmp6] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pavgb %[ftmp5], %[ftmp5], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" - "sdc1 %[ftmp4], 0x08(%[dst]) \n\t" - "gssdxc1 %[ftmp5], 0x08(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) + MMI_SDC1(%[ftmp4], %[dst], 0x08) + MMI_SDXC1(%[ftmp5], %[dst], %[dst_stride], 0x08) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -565,6 +516,8 @@ inline void ff_put_pixels16_l2_8_mmi(uint8_t *dst, const uint8_t *src1, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), @@ -583,60 +536,50 @@ inline void ff_avg_pixels4_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[6]; mips_reg addr[6]; - uint64_t low32; + DECLARE_VAR_LOW32; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" PTR_ADDU "%[addr3], %[src_stride2], %[src_stride2] \n\t" PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "uld %[low32], 0x00(%[src1]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[src2]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp0], %[src1], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_ULWC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" PTR_ADDU "%[addr5], %[dst], %[dst_stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x00(%[addr5]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_ULWC1(%[ftmp4], %[dst], 0x00) + MMI_ULWC1(%[ftmp5], %[addr5], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "swc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SWC1(%[ftmp0], %[dst], 0x00) + MMI_SWXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "uld %[low32], 0x00(%[src1]) \n\t" - "mtc1 %[low32], %[ftmp0] \n\t" - "uld %[low32], 0x00(%[addr0]) \n\t" - "mtc1 %[low32], %[ftmp1] \n\t" - "uld %[low32], 0x00(%[src2]) \n\t" - "mtc1 %[low32], %[ftmp2] \n\t" + MMI_ULWC1(%[ftmp0], %[src1], 0x00) + MMI_ULWC1(%[ftmp1], %[addr0], 0x00) + MMI_ULWC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "uld %[low32], 0x00(%[addr1]) \n\t" - "mtc1 %[low32], %[ftmp3] \n\t" + MMI_ULWC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" PTR_ADDU "%[addr5], %[dst], %[dst_stride] \n\t" - "uld %[low32], 0x00(%[dst]) \n\t" - "mtc1 %[low32], %[ftmp4] \n\t" - "uld %[low32], 0x00(%[addr5]) \n\t" - "mtc1 %[low32], %[ftmp5] \n\t" + MMI_ULWC1(%[ftmp4], %[dst], 0x00) + MMI_ULWC1(%[ftmp5], %[addr5], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "swc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gsswxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SWC1(%[ftmp0], %[dst], 0x00) + MMI_SWXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -645,10 +588,11 @@ inline void ff_avg_pixels4_l2_8_mmi(uint8_t *dst, const uint8_t *src1, : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), + RESTRICT_ASM_LOW32 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), [addr5]"=&r"(addr[5]), - [low32]"=&r"(low32), [dst]"+&r"(dst), [src1]"+&r"(src1), [src2]"+&r"(src2), [h]"+&r"(h) : [dst_stride]"r"((mips_reg)dst_stride), @@ -664,59 +608,50 @@ inline void ff_avg_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[6]; mips_reg addr[6]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" PTR_ADDU "%[addr3], %[src_stride2], %[src_stride2] \n\t" PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp2], %[src2], 0x00) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" PTR_ADDU "%[addr5], %[dst], %[dst_stride] \n\t" - "gsldlc1 %[ftmp4], 0x07(%[dst]) \n\t" - "gsldrc1 %[ftmp4], 0x00(%[dst]) \n\t" - "gsldlc1 %[ftmp5], 0x07(%[addr5]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[addr5]) \n\t" + MMI_ULDC1(%[ftmp4], %[dst], 0x00) + MMI_ULDC1(%[ftmp5], %[addr5], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp2], %[src2], 0x00) + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" PTR_ADDU "%[addr5], %[dst], %[dst_stride] \n\t" - "gsldlc1 %[ftmp4], 0x07(%[dst]) \n\t" - "gsldrc1 %[ftmp4], 0x00(%[dst]) \n\t" - "gsldlc1 %[ftmp5], 0x07(%[addr5]) \n\t" - "gsldrc1 %[ftmp5], 0x00(%[addr5]) \n\t" + MMI_ULDC1(%[ftmp4], %[dst], 0x00) + MMI_ULDC1(%[ftmp5], %[addr5], 0x00) "pavgb %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "pavgb %[ftmp1], %[ftmp1], %[ftmp5] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -725,6 +660,8 @@ inline void ff_avg_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), [addr5]"=&r"(addr[5]), @@ -795,24 +732,23 @@ inline void ff_put_no_rnd_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, { double ftmp[5]; mips_reg addr[5]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( "pcmpeqb %[ftmp4], %[ftmp4], %[ftmp4] \n\t" PTR_ADDU "%[addr2], %[src_stride1], %[src_stride1] \n\t" PTR_ADDU "%[addr3], %[src_stride2], %[src_stride2] \n\t" PTR_ADDU "%[addr4], %[dst_stride], %[dst_stride] \n\t" + "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp4] \n\t" "xor %[ftmp2], %[ftmp2], %[ftmp4] \n\t" @@ -821,22 +757,18 @@ inline void ff_put_no_rnd_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp4] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" + MMI_ULDC1(%[ftmp0], %[src1], 0x00) PTR_ADDU "%[addr0], %[src1], %[src_stride1] \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" + MMI_ULDC1(%[ftmp1], %[addr0], 0x00) + MMI_ULDC1(%[ftmp2], %[src2], 0x00) PTR_ADDU "%[addr1], %[src2], %[src_stride2] \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" - "gsldlc1 %[ftmp3], 0x07(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp3], %[addr1], 0x00) PTR_ADDU "%[src1], %[src1], %[addr2] \n\t" - "gsldrc1 %[ftmp3], 0x00(%[addr1]) \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp4] \n\t" "xor %[ftmp2], %[ftmp2], %[ftmp4] \n\t" @@ -845,8 +777,8 @@ inline void ff_put_no_rnd_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp4] \n\t" - "sdc1 %[ftmp0], 0x00(%[dst]) \n\t" - "gssdxc1 %[ftmp1], 0x00(%[dst], %[dst_stride]) \n\t" + MMI_SDC1(%[ftmp0], %[dst], 0x00) + MMI_SDXC1(%[ftmp1], %[dst], %[dst_stride], 0x00) PTR_ADDU "%[src2], %[src2], %[addr3] \n\t" PTR_ADDU "%[dst], %[dst], %[addr4] \n\t" @@ -855,6 +787,8 @@ inline void ff_put_no_rnd_pixels8_l2_8_mmi(uint8_t *dst, const uint8_t *src1, : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [addr2]"=&r"(addr[2]), [addr3]"=&r"(addr[3]), [addr4]"=&r"(addr[4]), @@ -981,6 +915,8 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, #if 1 double ftmp[10]; mips_reg addr[2]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" @@ -993,11 +929,9 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, "psllh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" "dli %[addr0], 0x02 \n\t" - "gsldlc1 %[ftmp0], 0x07(%[pixels]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[pixels]) \n\t" "dmtc1 %[addr0], %[ftmp9] \n\t" - "gsldlc1 %[ftmp4], 0x08(%[pixels]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[pixels]) \n\t" + MMI_ULDC1(%[ftmp0], %[pixels], 0x00) + MMI_ULDC1(%[ftmp4], %[pixels], 0x01) "mov.d %[ftmp1], %[ftmp0] \n\t" "mov.d %[ftmp5], %[ftmp4] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -1009,12 +943,11 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, "xor %[addr0], %[addr0], %[addr0] \n\t" PTR_ADDU "%[pixels], %[pixels], %[line_size] \n\t" ".p2align 3 \n\t" + "1: \n\t" PTR_ADDU "%[addr1], %[pixels], %[addr0] \n\t" - "gsldlc1 %[ftmp0], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp2], 0x08(%[addr1]) \n\t" - "gsldrc1 %[ftmp2], 0x01(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp0], %[addr1], 0x00) + MMI_ULDC1(%[ftmp2], %[addr1], 0x01) "mov.d %[ftmp1], %[ftmp0] \n\t" "mov.d %[ftmp3], %[ftmp2] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp7] \n\t" @@ -1030,13 +963,11 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, "psrlh %[ftmp4], %[ftmp4], %[ftmp9] \n\t" "psrlh %[ftmp5], %[ftmp5], %[ftmp9] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp5] \n\t" - "gssdxc1 %[ftmp4], 0x00(%[block], %[addr0]) \n\t" + MMI_SDXC1(%[ftmp4], %[block], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[line_size] \n\t" PTR_ADDU "%[addr1], %[pixels], %[addr0] \n\t" - "gsldlc1 %[ftmp2], 0x07(%[addr1]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[addr1]) \n\t" - "gsldlc1 %[ftmp4], 0x08(%[addr1]) \n\t" - "gsldrc1 %[ftmp4], 0x01(%[addr1]) \n\t" + MMI_ULDC1(%[ftmp2], %[addr1], 0x00) + MMI_ULDC1(%[ftmp4], %[addr1], 0x01) "mov.d %[ftmp3], %[ftmp2] \n\t" "mov.d %[ftmp5], %[ftmp4] \n\t" "punpcklbh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" @@ -1052,7 +983,7 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, "psrlh %[ftmp0], %[ftmp0], %[ftmp9] \n\t" "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp1] \n\t" - "gssdxc1 %[ftmp0], 0x00(%[block], %[addr0]) \n\t" + MMI_SDXC1(%[ftmp0], %[block], %[addr0], 0x00) PTR_ADDU "%[addr0], %[addr0], %[line_size] \n\t" PTR_ADDU "%[h], %[h], -0x02 \n\t" "bnez %[h], 1b \n\t" @@ -1061,6 +992,8 @@ void ff_put_pixels8_xy2_8_mmi(uint8_t *block, const uint8_t *pixels, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [addr1]"=&r"(addr[1]), [h]"+&r"(h), [pixels]"+&r"(pixels) : [block]"r"(block), [line_size]"r"((mips_reg)line_size) diff --git a/libavcodec/mips/idctdsp_mmi.c b/libavcodec/mips/idctdsp_mmi.c index 24beb62..b797965 100644 --- a/libavcodec/mips/idctdsp_mmi.c +++ b/libavcodec/mips/idctdsp_mmi.c @@ -23,36 +23,40 @@ #include "idctdsp_mips.h" #include "constants.h" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_put_pixels_clamped_mmi(const int16_t *block, uint8_t *av_restrict pixels, ptrdiff_t line_size) { double ftmp[8]; mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( - "ldc1 %[ftmp0], 0x00(%[block]) \n\t" - "ldc1 %[ftmp1], 0x08(%[block]) \n\t" - "ldc1 %[ftmp2], 0x10(%[block]) \n\t" - "ldc1 %[ftmp3], 0x18(%[block]) \n\t" - "ldc1 %[ftmp4], 0x20(%[block]) \n\t" - "ldc1 %[ftmp5], 0x28(%[block]) \n\t" - "ldc1 %[ftmp6], 0x30(%[block]) \n\t" - "ldc1 %[ftmp7], 0x38(%[block]) \n\t" + MMI_LDC1(%[ftmp0], %[block], 0x00) + MMI_LDC1(%[ftmp1], %[block], 0x08) + MMI_LDC1(%[ftmp2], %[block], 0x10) + MMI_LDC1(%[ftmp3], %[block], 0x18) + MMI_LDC1(%[ftmp4], %[block], 0x20) + MMI_LDC1(%[ftmp5], %[block], 0x28) + MMI_LDC1(%[ftmp6], %[block], 0x30) + MMI_LDC1(%[ftmp7], %[block], 0x38) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp1] \n\t" "packushb %[ftmp2], %[ftmp2], %[ftmp3] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp5] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" - "gssdxc1 %[ftmp4], 0x00(%[addr0], %[line_size]) \n\t" - "gssdxc1 %[ftmp6], 0x00(%[pixels], %[line_sizex3]) \n\t" + MMI_SDC1(%[ftmp0], %[pixels], 0x00) + MMI_SDC1(%[ftmp2], %[addr0], 0x00) + MMI_SDXC1(%[ftmp4], %[addr0], %[line_size], 0x00) + MMI_SDXC1(%[ftmp6], %[pixels], %[line_sizex3], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [pixels]"+&r"(pixels) : [line_size]"r"((mips_reg)line_size), @@ -65,27 +69,29 @@ void ff_put_pixels_clamped_mmi(const int16_t *block, block += 32; __asm__ volatile ( - "ldc1 %[ftmp0], 0x00(%[block]) \n\t" - "ldc1 %[ftmp1], 0x08(%[block]) \n\t" - "ldc1 %[ftmp2], 0x10(%[block]) \n\t" - "ldc1 %[ftmp3], 0x18(%[block]) \n\t" - "ldc1 %[ftmp4], 0x20(%[block]) \n\t" - "ldc1 %[ftmp5], 0x28(%[block]) \n\t" - "ldc1 %[ftmp6], 0x30(%[block]) \n\t" - "ldc1 %[ftmp7], 0x38(%[block]) \n\t" + MMI_LDC1(%[ftmp0], %[block], 0x00) + MMI_LDC1(%[ftmp1], %[block], 0x08) + MMI_LDC1(%[ftmp2], %[block], 0x10) + MMI_LDC1(%[ftmp3], %[block], 0x18) + MMI_LDC1(%[ftmp4], %[block], 0x20) + MMI_LDC1(%[ftmp5], %[block], 0x28) + MMI_LDC1(%[ftmp6], %[block], 0x30) + MMI_LDC1(%[ftmp7], %[block], 0x38) PTR_ADDU "%[addr0], %[pixels], %[line_size] \n\t" "packushb %[ftmp0], %[ftmp0], %[ftmp1] \n\t" "packushb %[ftmp2], %[ftmp2], %[ftmp3] \n\t" "packushb %[ftmp4], %[ftmp4], %[ftmp5] \n\t" "packushb %[ftmp6], %[ftmp6], %[ftmp7] \n\t" - "sdc1 %[ftmp0], 0x00(%[pixels]) \n\t" - "sdc1 %[ftmp2], 0x00(%[addr0]) \n\t" - "gssdxc1 %[ftmp4], 0x00(%[addr0], %[line_size]) \n\t" - "gssdxc1 %[ftmp6], 0x00(%[pixels], %[line_sizex3]) \n\t" + MMI_SDC1(%[ftmp0], %[pixels], 0x00) + MMI_SDC1(%[ftmp2], %[addr0], 0x00) + MMI_SDXC1(%[ftmp4], %[addr0], %[line_size], 0x00) + MMI_SDXC1(%[ftmp6], %[pixels], %[line_sizex3], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [pixels]"+&r"(pixels) : [line_size]"r"((mips_reg)line_size), @@ -102,56 +108,60 @@ void ff_put_signed_pixels_clamped_mmi(const int16_t *block, int64_t line_skip3 = 0; double ftmp[5]; mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( PTR_ADDU "%[line_skip3], %[line_skip], %[line_skip] \n\t" - "ldc1 %[ftmp1], 0x00(%[block]) \n\t" - "ldc1 %[ftmp0], 0x08(%[block]) \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x00) + MMI_LDC1(%[ftmp0], %[block], 0x08) "packsshb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "ldc1 %[ftmp2], 0x10(%[block]) \n\t" - "ldc1 %[ftmp0], 0x18(%[block]) \n\t" + MMI_LDC1(%[ftmp2], %[block], 0x10) + MMI_LDC1(%[ftmp0], %[block], 0x18) "packsshb %[ftmp2], %[ftmp2], %[ftmp0] \n\t" - "ldc1 %[ftmp3], 0x20(%[block]) \n\t" - "ldc1 %[ftmp0], 0x28(%[block]) \n\t" + MMI_LDC1(%[ftmp3], %[block], 0x20) + MMI_LDC1(%[ftmp0], %[block], 0x28) "packsshb %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "ldc1 %[ftmp4], 48(%[block]) \n\t" - "ldc1 %[ftmp0], 56(%[block]) \n\t" + MMI_LDC1(%[ftmp4], %[block], 0x30) + MMI_LDC1(%[ftmp0], %[block], 0x38) "packsshb %[ftmp4], %[ftmp4], %[ftmp0] \n\t" "paddb %[ftmp1], %[ftmp1], %[ff_pb_80] \n\t" "paddb %[ftmp2], %[ftmp2], %[ff_pb_80] \n\t" "paddb %[ftmp3], %[ftmp3], %[ff_pb_80] \n\t" "paddb %[ftmp4], %[ftmp4], %[ff_pb_80] \n\t" - "sdc1 %[ftmp1], 0x00(%[pixels]) \n\t" - "gssdxc1 %[ftmp2], 0x00(%[pixels], %[line_skip]) \n\t" - "gssdxc1 %[ftmp3], 0x00(%[pixels], %[line_skip3]) \n\t" + MMI_SDC1(%[ftmp1], %[pixels], 0x00) + MMI_SDXC1(%[ftmp2], %[pixels], %[line_skip], 0x00) + MMI_SDXC1(%[ftmp3], %[pixels], %[line_skip3], 0x00) PTR_ADDU "%[line_skip3], %[line_skip3], %[line_skip] \n\t" - "gssdxc1 %[ftmp4], 0x00(%[pixels], %[line_skip3]) \n\t" + MMI_SDXC1(%[ftmp4], %[pixels], %[line_skip3], 0x00) PTR_ADDU "%[addr0], %[line_skip3], %[line_skip] \n\t" PTR_ADDU "%[pixels], %[pixels], %[addr0] \n\t" - "ldc1 %[ftmp1], 0x40(%[block]) \n\t" - "ldc1 %[ftmp0], 0x48(%[block]) \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x40) + MMI_LDC1(%[ftmp0], %[block], 0x48) "packsshb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" - "ldc1 %[ftmp2], 0x50(%[block]) \n\t" - "ldc1 %[ftmp0], 0x58(%[block]) \n\t" + MMI_LDC1(%[ftmp2], %[block], 0x50) + MMI_LDC1(%[ftmp0], %[block], 0x58) "packsshb %[ftmp2], %[ftmp2], %[ftmp0] \n\t" - "ldc1 %[ftmp3], 0x60(%[block]) \n\t" - "ldc1 %[ftmp0], 0x68(%[block]) \n\t" + MMI_LDC1(%[ftmp3], %[block], 0x60) + MMI_LDC1(%[ftmp0], %[block], 0x68) "packsshb %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "ldc1 %[ftmp4], 0x70(%[block]) \n\t" - "ldc1 %[ftmp0], 0x78(%[block]) \n\t" + MMI_LDC1(%[ftmp4], %[block], 0x70) + MMI_LDC1(%[ftmp0], %[block], 0x78) "packsshb %[ftmp4], %[ftmp4], %[ftmp0] \n\t" "paddb %[ftmp1], %[ftmp1], %[ff_pb_80] \n\t" "paddb %[ftmp2], %[ftmp2], %[ff_pb_80] \n\t" "paddb %[ftmp3], %[ftmp3], %[ff_pb_80] \n\t" "paddb %[ftmp4], %[ftmp4], %[ff_pb_80] \n\t" - "sdc1 %[ftmp1], 0x00(%[pixels]) \n\t" - "gssdxc1 %[ftmp2], 0x00(%[pixels], %[line_skip]) \n\t" + MMI_SDC1(%[ftmp1], %[pixels], 0x00) + MMI_SDXC1(%[ftmp2], %[pixels], %[line_skip], 0x00) PTR_ADDU "%[addr0], %[line_skip], %[line_skip] \n\t" - "gssdxc1 %[ftmp3], 0x00(%[pixels], %[addr0]) \n\t" - "gssdxc1 %[ftmp4], 0x00(%[pixels], %[line_skip3]) \n\t" + MMI_SDXC1(%[ftmp3], %[pixels], %[addr0], 0x00) + MMI_SDXC1(%[ftmp4], %[pixels], %[line_skip3], 0x00) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]), [pixels]"+&r"(pixels), [line_skip3]"+&r"(line_skip3) : [block]"r"(block), @@ -166,17 +176,20 @@ void ff_add_pixels_clamped_mmi(const int16_t *block, { double ftmp[8]; uint64_t tmp[1]; + mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( "li %[tmp0], 0x04 \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "1: \n\t" - "ldc1 %[ftmp1], 0x00(%[block]) \n\t" - "ldc1 %[ftmp2], 0x08(%[block]) \n\t" - "ldc1 %[ftmp3], 0x10(%[block]) \n\t" - "ldc1 %[ftmp4], 0x18(%[block]) \n\t" - "ldc1 %[ftmp5], 0x00(%[pixels]) \n\t" - "gsldxc1 %[ftmp6], 0x00(%[pixels], %[line_size]) \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x00) + MMI_LDC1(%[ftmp2], %[block], 0x08) + MMI_LDC1(%[ftmp3], %[block], 0x10) + MMI_LDC1(%[ftmp4], %[block], 0x18) + MMI_LDC1(%[ftmp5], %[pixels], 0x00) + MMI_LDXC1(%[ftmp6], %[pixels], %[line_size], 0x00) "mov.d %[ftmp7], %[ftmp5] \n\t" "punpcklbh %[ftmp5], %[ftmp5], %[ftmp0] \n\t" "punpckhbh %[ftmp7], %[ftmp7], %[ftmp0] \n\t" @@ -189,8 +202,8 @@ void ff_add_pixels_clamped_mmi(const int16_t *block, "paddh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "packushb %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "sdc1 %[ftmp1], 0x00(%[pixels]) \n\t" - "gssdxc1 %[ftmp3], 0x00(%[pixels], %[line_size]) \n\t" + MMI_SDC1(%[ftmp1], %[pixels], 0x00) + MMI_SDXC1(%[ftmp3], %[pixels], %[line_size], 0x00) "addi %[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDIU "%[block], %[block], 0x20 \n\t" PTR_ADDU "%[pixels], %[pixels], %[line_size] \n\t" @@ -201,6 +214,9 @@ void ff_add_pixels_clamped_mmi(const int16_t *block, [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT + [addr0]"=&r"(addr[0]), [pixels]"+&r"(pixels), [block]"+&r"(block) : [line_size]"r"((mips_reg)line_size) : "memory" diff --git a/libavcodec/mips/mpegvideo_mmi.c b/libavcodec/mips/mpegvideo_mmi.c index 450a18c..18058e4 100644 --- a/libavcodec/mips/mpegvideo_mmi.c +++ b/libavcodec/mips/mpegvideo_mmi.c @@ -23,7 +23,7 @@ */ #include "mpegvideo_mips.h" -#include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_dct_unquantize_h263_intra_mmi(MpegEncContext *s, int16_t *block, int n, int qscale) @@ -31,6 +31,7 @@ void ff_dct_unquantize_h263_intra_mmi(MpegEncContext *s, int16_t *block, int64_t level, qmul, qadd, nCoeffs; double ftmp[6]; mips_reg addr[1]; + DECLARE_VAR_ALL64; qmul = qscale << 1; av_assert2(s->block_last_index[n]>=0 || s->h263_aic); @@ -60,12 +61,11 @@ void ff_dct_unquantize_h263_intra_mmi(MpegEncContext *s, int16_t *block, "psubh %[ftmp0], %[ftmp0], %[qadd] \n\t" "xor %[ftmp5], %[ftmp5], %[ftmp5] \n\t" ".p2align 4 \n\t" + "1: \n\t" PTR_ADDU "%[addr0], %[block], %[nCoeffs] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp2], 0x08(%[addr0]) \n\t" + MMI_LDC1(%[ftmp1], %[addr0], 0x00) + MMI_LDC1(%[ftmp2], %[addr0], 0x08) "mov.d %[ftmp3], %[ftmp1] \n\t" "mov.d %[ftmp4], %[ftmp2] \n\t" "pmullh %[ftmp1], %[ftmp1], %[qmul] \n\t" @@ -83,14 +83,13 @@ void ff_dct_unquantize_h263_intra_mmi(MpegEncContext *s, int16_t *block, "pandn %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pandn %[ftmp2], %[ftmp2], %[ftmp4] \n\t" PTR_ADDIU "%[nCoeffs], %[nCoeffs], 0x10 \n\t" - "gssdlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gssdlc1 %[ftmp2], 0x0f(%[addr0]) \n\t" - "gssdrc1 %[ftmp2], 0x08(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) + MMI_SDC1(%[ftmp2], %[addr0], 0x08) "blez %[nCoeffs], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]) : [block]"r"((mips_reg)(block+nCoeffs)), [nCoeffs]"r"((mips_reg)(2*(-nCoeffs))), @@ -107,6 +106,7 @@ void ff_dct_unquantize_h263_inter_mmi(MpegEncContext *s, int16_t *block, int64_t qmul, qadd, nCoeffs; double ftmp[6]; mips_reg addr[1]; + DECLARE_VAR_ALL64; qmul = qscale << 1; qadd = (qscale - 1) | 1; @@ -124,10 +124,8 @@ void ff_dct_unquantize_h263_inter_mmi(MpegEncContext *s, int16_t *block, ".p2align 4 \n\t" "1: \n\t" PTR_ADDU "%[addr0], %[block], %[nCoeffs] \n\t" - "gsldlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gsldlc1 %[ftmp2], 0x0f(%[addr0]) \n\t" - "gsldrc1 %[ftmp2], 0x08(%[addr0]) \n\t" + MMI_LDC1(%[ftmp1], %[addr0], 0x00) + MMI_LDC1(%[ftmp2], %[addr0], 0x08) "mov.d %[ftmp3], %[ftmp1] \n\t" "mov.d %[ftmp4], %[ftmp2] \n\t" "pmullh %[ftmp1], %[ftmp1], %[qmul] \n\t" @@ -145,14 +143,13 @@ void ff_dct_unquantize_h263_inter_mmi(MpegEncContext *s, int16_t *block, "pandn %[ftmp1], %[ftmp1], %[ftmp3] \n\t" "pandn %[ftmp2], %[ftmp2], %[ftmp4] \n\t" PTR_ADDIU "%[nCoeffs], %[nCoeffs], 0x10 \n\t" - "gssdlc1 %[ftmp1], 0x07(%[addr0]) \n\t" - "gssdrc1 %[ftmp1], 0x00(%[addr0]) \n\t" - "gssdlc1 %[ftmp2], 0x0f(%[addr0]) \n\t" - "gssdrc1 %[ftmp2], 0x08(%[addr0]) \n\t" + MMI_SDC1(%[ftmp1], %[addr0], 0x00) + MMI_SDC1(%[ftmp2], %[addr0], 0x08) "blez %[nCoeffs], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]) : [block]"r"((mips_reg)(block+nCoeffs)), [nCoeffs]"r"((mips_reg)(2*(-nCoeffs))), @@ -170,6 +167,8 @@ void ff_dct_unquantize_mpeg1_intra_mmi(MpegEncContext *s, int16_t *block, double ftmp[10]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; av_assert2(s->block_last_index[n]>=0); nCoeffs = s->intra_scantable.raster_end[s->block_last_index[n]] + 1; @@ -192,13 +191,14 @@ void ff_dct_unquantize_mpeg1_intra_mmi(MpegEncContext *s, int16_t *block, "packsswh %[ftmp1], %[ftmp1], %[ftmp1] \n\t" "or %[addr0], %[nCoeffs], $0 \n\t" ".p2align 4 \n\t" + "1: \n\t" - "gsldxc1 %[ftmp2], 0x00(%[addr0], %[block]) \n\t" - "gsldxc1 %[ftmp3], 0x08(%[addr0], %[block]) \n\t" + MMI_LDXC1(%[ftmp2], %[addr0], %[block], 0x00) + MMI_LDXC1(%[ftmp3], %[addr0], %[block], 0x08) "mov.d %[ftmp4], %[ftmp2] \n\t" "mov.d %[ftmp5], %[ftmp3] \n\t" - "gsldxc1 %[ftmp6], 0x00(%[addr0], %[quant]) \n\t" - "gsldxc1 %[ftmp7], 0x08(%[addr0], %[quant]) \n\t" + MMI_LDXC1(%[ftmp6], %[addr0], %[quant], 0x00) + MMI_LDXC1(%[ftmp7], %[addr0], %[quant], 0x08) "pmullh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "pmullh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" "xor %[ftmp8], %[ftmp8], %[ftmp8] \n\t" @@ -229,8 +229,8 @@ void ff_dct_unquantize_mpeg1_intra_mmi(MpegEncContext *s, int16_t *block, "psubh %[ftmp3], %[ftmp3], %[ftmp9] \n\t" "pandn %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "pandn %[ftmp7], %[ftmp7], %[ftmp3] \n\t" - "gssdxc1 %[ftmp6], 0x00(%[addr0], %[block]) \n\t" - "gssdxc1 %[ftmp7], 0x08(%[addr0], %[block]) \n\t" + MMI_SDXC1(%[ftmp6], %[addr0], %[block], 0x00) + MMI_SDXC1(%[ftmp7], %[addr0], %[block], 0x08) PTR_ADDIU "%[addr0], %[addr0], 0x10 \n\t" "bltz %[addr0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), @@ -239,6 +239,8 @@ void ff_dct_unquantize_mpeg1_intra_mmi(MpegEncContext *s, int16_t *block, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]) : [block]"r"((mips_reg)(block+nCoeffs)), [quant]"r"((mips_reg)(quant_matrix+nCoeffs)), @@ -258,6 +260,8 @@ void ff_dct_unquantize_mpeg1_inter_mmi(MpegEncContext *s, int16_t *block, double ftmp[10]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; av_assert2(s->block_last_index[n] >= 0); nCoeffs = s->intra_scantable.raster_end[s->block_last_index[n]] + 1; @@ -273,13 +277,14 @@ void ff_dct_unquantize_mpeg1_inter_mmi(MpegEncContext *s, int16_t *block, "packsswh %[ftmp1], %[ftmp1], %[ftmp1] \n\t" "or %[addr0], %[nCoeffs], $0 \n\t" ".p2align 4 \n\t" + "1: \n\t" - "gsldxc1 %[ftmp2], 0x00(%[addr0], %[block]) \n\t" - "gsldxc1 %[ftmp3], 0x08(%[addr0], %[block]) \n\t" + MMI_LDXC1(%[ftmp2], %[addr0], %[block], 0x00) + MMI_LDXC1(%[ftmp3], %[addr0], %[block], 0x08) "mov.d %[ftmp4], %[ftmp2] \n\t" "mov.d %[ftmp5], %[ftmp3] \n\t" - "gsldxc1 %[ftmp6], 0x00(%[addr0], %[quant]) \n\t" - "gsldxc1 %[ftmp7], 0x08(%[addr0], %[quant]) \n\t" + MMI_LDXC1(%[ftmp6], %[addr0], %[quant], 0x00) + MMI_LDXC1(%[ftmp7], %[addr0], %[quant], 0x08) "pmullh %[ftmp6], %[ftmp6], %[ftmp1] \n\t" "pmullh %[ftmp7], %[ftmp7], %[ftmp1] \n\t" "xor %[ftmp8], %[ftmp8], %[ftmp8] \n\t" @@ -314,8 +319,8 @@ void ff_dct_unquantize_mpeg1_inter_mmi(MpegEncContext *s, int16_t *block, "psubh %[ftmp3], %[ftmp3], %[ftmp9] \n\t" "pandn %[ftmp6], %[ftmp6], %[ftmp2] \n\t" "pandn %[ftmp7], %[ftmp7], %[ftmp3] \n\t" - "gssdxc1 %[ftmp6], 0x00(%[addr0], %[block]) \n\t" - "gssdxc1 %[ftmp7], 0x08(%[addr0], %[block]) \n\t" + MMI_SDXC1(%[ftmp6], %[addr0], %[block], 0x00) + MMI_SDXC1(%[ftmp7], %[addr0], %[block], 0x08) PTR_ADDIU "%[addr0], %[addr0], 0x10 \n\t" "bltz %[addr0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), @@ -324,6 +329,8 @@ void ff_dct_unquantize_mpeg1_inter_mmi(MpegEncContext *s, int16_t *block, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]) : [block]"r"((mips_reg)(block+nCoeffs)), [quant]"r"((mips_reg)(quant_matrix+nCoeffs)), @@ -342,6 +349,8 @@ void ff_dct_unquantize_mpeg2_intra_mmi(MpegEncContext *s, int16_t *block, double ftmp[10]; uint64_t tmp[1]; mips_reg addr[1]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; assert(s->block_last_index[n]>=0); @@ -367,13 +376,14 @@ void ff_dct_unquantize_mpeg2_intra_mmi(MpegEncContext *s, int16_t *block, "packsswh %[ftmp9], %[ftmp9], %[ftmp9] \n\t" "or %[addr0], %[nCoeffs], $0 \n\t" ".p2align 4 \n\t" + "1: \n\t" - "gsldxc1 %[ftmp1], 0x00(%[addr0], %[block]) \n\t" - "gsldxc1 %[ftmp2], 0x08(%[addr0], %[block]) \n\t" + MMI_LDXC1(%[ftmp1], %[addr0], %[block], 0x00) + MMI_LDXC1(%[ftmp2], %[addr0], %[block], 0x08) "mov.d %[ftmp3], %[ftmp1] \n\t" "mov.d %[ftmp4], %[ftmp2] \n\t" - "gsldxc1 %[ftmp5], 0x00(%[addr0], %[quant]) \n\t" - "gsldxc1 %[ftmp6], 0x00(%[addr0], %[quant]) \n\t" + MMI_LDXC1(%[ftmp5], %[addr0], %[quant], 0x00) + MMI_LDXC1(%[ftmp6], %[addr0], %[quant], 0x08) "pmullh %[ftmp5], %[ftmp5], %[ftmp9] \n\t" "pmullh %[ftmp6], %[ftmp6], %[ftmp9] \n\t" "xor %[ftmp7], %[ftmp7], %[ftmp7] \n\t" @@ -401,8 +411,8 @@ void ff_dct_unquantize_mpeg2_intra_mmi(MpegEncContext *s, int16_t *block, "pandn %[ftmp5], %[ftmp5], %[ftmp1] \n\t" "pandn %[ftmp6], %[ftmp6], %[ftmp2] \n\t" PTR_ADDIU "%[addr0], %[addr0], 0x10 \n\t" - "gssdxc1 %[ftmp5], 0x00(%[addr0], %[block]) \n\t" - "gssdxc1 %[ftmp6], 0x08(%[addr0], %[block]) \n\t" + MMI_SDXC1(%[ftmp5], %[addr0], %[block], 0x00) + MMI_SDXC1(%[ftmp6], %[addr0], %[block], 0x08) "blez %[addr0], 1b \n\t" : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), @@ -410,6 +420,8 @@ void ff_dct_unquantize_mpeg2_intra_mmi(MpegEncContext *s, int16_t *block, [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [addr0]"=&r"(addr[0]) : [block]"r"((mips_reg)(block+nCoeffs)), [quant]"r"((mips_reg)(quant_matrix+nCoeffs)), @@ -428,15 +440,16 @@ void ff_denoise_dct_mmi(MpegEncContext *s, int16_t *block) uint16_t *offset = s->dct_offset[intra]; double ftmp[8]; mips_reg addr[1]; + DECLARE_VAR_ALL64; s->dct_count[intra]++; __asm__ volatile( "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" "1: \n\t" - "ldc1 %[ftmp1], 0x00(%[block]) \n\t" + MMI_LDC1(%[ftmp1], %[block], 0x00) "xor %[ftmp2], %[ftmp2], %[ftmp2] \n\t" - "ldc1 %[ftmp3], 0x08(%[block]) \n\t" + MMI_LDC1(%[ftmp3], %[block], 0x08) "xor %[ftmp4], %[ftmp4], %[ftmp4] \n\t" "pcmpgth %[ftmp2], %[ftmp2], %[ftmp1] \n\t" "pcmpgth %[ftmp4], %[ftmp4], %[ftmp3] \n\t" @@ -444,36 +457,36 @@ void ff_denoise_dct_mmi(MpegEncContext *s, int16_t *block) "xor %[ftmp3], %[ftmp3], %[ftmp4] \n\t" "psubh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "ldc1 %[ftmp6], 0x00(%[offset]) \n\t" + MMI_LDC1(%[ftmp6], %[offset], 0x00) "mov.d %[ftmp5], %[ftmp1] \n\t" "psubush %[ftmp1], %[ftmp1], %[ftmp6] \n\t" - "ldc1 %[ftmp6], 0x08(%[offset]) \n\t" + MMI_LDC1(%[ftmp6], %[offset], 0x08) "mov.d %[ftmp7], %[ftmp3] \n\t" "psubush %[ftmp3], %[ftmp3], %[ftmp6] \n\t" "xor %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "xor %[ftmp3], %[ftmp3], %[ftmp4] \n\t" "psubh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" "psubh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" - "sdc1 %[ftmp1], 0x00(%[block]) \n\t" - "sdc1 %[ftmp3], 0x08(%[block]) \n\t" + MMI_SDC1(%[ftmp1], %[block], 0x00) + MMI_SDC1(%[ftmp3], %[block], 0x08) "mov.d %[ftmp1], %[ftmp5] \n\t" "mov.d %[ftmp3], %[ftmp7] \n\t" "punpcklhw %[ftmp5], %[ftmp5], %[ftmp0] \n\t" "punpckhhw %[ftmp1], %[ftmp1], %[ftmp0] \n\t" "punpcklhw %[ftmp7], %[ftmp7], %[ftmp0] \n\t" "punpckhhw %[ftmp3], %[ftmp3], %[ftmp0] \n\t" - "ldc1 %[ftmp2], 0x00(%[sum]) \n\t" + MMI_LDC1(%[ftmp2], %[sum], 0x00) "paddw %[ftmp5], %[ftmp5], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x08(%[sum]) \n\t" + MMI_LDC1(%[ftmp2], %[sum], 0x08) "paddw %[ftmp1], %[ftmp1], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x10(%[sum]) \n\t" + MMI_LDC1(%[ftmp2], %[sum], 0x10) "paddw %[ftmp7], %[ftmp7], %[ftmp2] \n\t" - "ldc1 %[ftmp2], 0x18(%[sum]) \n\t" + MMI_LDC1(%[ftmp2], %[sum], 0x18) "paddw %[ftmp3], %[ftmp3], %[ftmp2] \n\t" - "sdc1 %[ftmp5], 0x00(%[sum]) \n\t" - "sdc1 %[ftmp1], 0x08(%[sum]) \n\t" - "sdc1 %[ftmp7], 0x10(%[sum]) \n\t" - "sdc1 %[ftmp3], 0x18(%[sum]) \n\t" + MMI_SDC1(%[ftmp5], %[sum], 0x00) + MMI_SDC1(%[ftmp1], %[sum], 0x08) + MMI_SDC1(%[ftmp7], %[sum], 0x10) + MMI_SDC1(%[ftmp3], %[sum], 0x18) PTR_ADDIU "%[block], %[block], 0x10 \n\t" PTR_ADDIU "%[sum], %[sum], 0x20 \n\t" PTR_SUBU "%[addr0], %[block1], %[block] \n\t" @@ -483,6 +496,7 @@ void ff_denoise_dct_mmi(MpegEncContext *s, int16_t *block) [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), + RESTRICT_ASM_ALL64 [addr0]"=&r"(addr[0]), [block]"+&r"(block), [sum]"+&r"(sum), [offset]"+&r"(offset) diff --git a/libavcodec/mips/pixblockdsp_mmi.c b/libavcodec/mips/pixblockdsp_mmi.c index 3ff84c0..9f2eac3 100644 --- a/libavcodec/mips/pixblockdsp_mmi.c +++ b/libavcodec/mips/pixblockdsp_mmi.c @@ -23,34 +23,73 @@ #include "pixblockdsp_mips.h" #include "libavutil/mips/asmdefs.h" +#include "libavutil/mips/mmiutils.h" void ff_get_pixels_8_mmi(int16_t *av_restrict block, const uint8_t *pixels, ptrdiff_t line_size) { - double ftmp[6]; - mips_reg tmp[2]; + double ftmp[7]; + DECLARE_VAR_ALL64; + DECLARE_VAR_ADDRT; __asm__ volatile ( - "li %[tmp1], 0x08 \n\t" - "move %[tmp0], $0 \n\t" "xor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" - "1: \n\t" - "gsldlc1 %[ftmp1], 0x07(%[pixels]) \n\t" - "gsldrc1 %[ftmp1], 0x00(%[pixels]) \n\t" - "punpcklbh %[ftmp2], %[ftmp1], %[ftmp0] \n\t" - "punpckhbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" - "gssdxc1 %[ftmp2], 0x00(%[block], %[tmp0]) \n\t" - "gssdxc1 %[ftmp5], 0x08(%[block], %[tmp0]) \n\t" - PTR_ADDI "%[tmp1], %[tmp1], -0x01 \n\t" - PTR_ADDIU "%[tmp0], %[tmp0], 0x10 \n\t" - PTR_ADDU "%[pixels], %[pixels], %[line_size] \n\t" - "bnez %[tmp1], 1b \n\t" + + MMI_LDC1(%[ftmp1], %[pixels], 0x00) + MMI_LDXC1(%[ftmp2], %[pixels], %[line_size], 0x00) + "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" + "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" + "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" + "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp3], %[block], 0x00) + MMI_SDC1(%[ftmp4], %[block], 0x08) + MMI_SDC1(%[ftmp5], %[block], 0x10) + MMI_SDC1(%[ftmp6], %[block], 0x18) + PTR_ADDU "%[pixels], %[pixels], %[line_size_x2] \n\t" + + MMI_LDC1(%[ftmp1], %[pixels], 0x00) + MMI_LDXC1(%[ftmp2], %[pixels], %[line_size], 0x00) + "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" + "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" + "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" + "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp3], %[block], 0x20) + MMI_SDC1(%[ftmp4], %[block], 0x28) + MMI_SDC1(%[ftmp5], %[block], 0x30) + MMI_SDC1(%[ftmp6], %[block], 0x38) + PTR_ADDU "%[pixels], %[pixels], %[line_size_x2] \n\t" + + MMI_LDC1(%[ftmp1], %[pixels], 0x00) + MMI_LDXC1(%[ftmp2], %[pixels], %[line_size], 0x00) + "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" + "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" + "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" + "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp3], %[block], 0x40) + MMI_SDC1(%[ftmp4], %[block], 0x48) + MMI_SDC1(%[ftmp5], %[block], 0x50) + MMI_SDC1(%[ftmp6], %[block], 0x58) + PTR_ADDU "%[pixels], %[pixels], %[line_size_x2] \n\t" + + MMI_LDC1(%[ftmp1], %[pixels], 0x00) + MMI_LDXC1(%[ftmp2], %[pixels], %[line_size], 0x00) + "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" + "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" + "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" + "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" + MMI_SDC1(%[ftmp3], %[block], 0x60) + MMI_SDC1(%[ftmp4], %[block], 0x68) + MMI_SDC1(%[ftmp5], %[block], 0x70) + MMI_SDC1(%[ftmp6], %[block], 0x78) : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), - [tmp0]"=&r"(tmp[0]), [tmp1]"=&r"(tmp[1]), + [ftmp6]"=&f"(ftmp[6]), + RESTRICT_ASM_ALL64 + RESTRICT_ASM_ADDRT [pixels]"+&r"(pixels) - : [block]"r"((mips_reg)block), [line_size]"r"((mips_reg)line_size) + : [block]"r"((mips_reg)block), [line_size]"r"((mips_reg)line_size), + [line_size_x2]"r"((mips_reg)(line_size<<1)) : "memory" ); } @@ -60,16 +99,15 @@ void ff_diff_pixels_mmi(int16_t *av_restrict block, const uint8_t *src1, { double ftmp[5]; mips_reg tmp[1]; + DECLARE_VAR_ALL64; __asm__ volatile ( "li %[tmp0], 0x08 \n\t" "xor %[ftmp4], %[ftmp4], %[ftmp4] \n\t" "1: \n\t" - "gsldlc1 %[ftmp0], 0x07(%[src1]) \n\t" - "gsldrc1 %[ftmp0], 0x00(%[src1]) \n\t" + MMI_LDC1(%[ftmp0], %[src1], 0x00) "or %[ftmp1], %[ftmp0], %[ftmp0] \n\t" - "gsldlc1 %[ftmp2], 0x07(%[src2]) \n\t" - "gsldrc1 %[ftmp2], 0x00(%[src2]) \n\t" + MMI_LDC1(%[ftmp2], %[src2], 0x00) "or %[ftmp3], %[ftmp2], %[ftmp2] \n\t" "punpcklbh %[ftmp0], %[ftmp0], %[ftmp4] \n\t" "punpckhbh %[ftmp1], %[ftmp1], %[ftmp4] \n\t" @@ -77,10 +115,8 @@ void ff_diff_pixels_mmi(int16_t *av_restrict block, const uint8_t *src1, "punpckhbh %[ftmp3], %[ftmp3], %[ftmp4] \n\t" "psubh %[ftmp0], %[ftmp0], %[ftmp2] \n\t" "psubh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" - "gssdlc1 %[ftmp0], 0x07(%[block]) \n\t" - "gssdrc1 %[ftmp0], 0x00(%[block]) \n\t" - "gssdlc1 %[ftmp1], 0x0f(%[block]) \n\t" - "gssdrc1 %[ftmp1], 0x08(%[block]) \n\t" + MMI_SDC1(%[ftmp0], %[block], 0x00) + MMI_SDC1(%[ftmp1], %[block], 0x08) PTR_ADDI "%[tmp0], %[tmp0], -0x01 \n\t" PTR_ADDIU "%[block], %[block], 0x10 \n\t" PTR_ADDU "%[src1], %[src1], %[stride] \n\t" @@ -90,6 +126,7 @@ void ff_diff_pixels_mmi(int16_t *av_restrict block, const uint8_t *src1, [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), [ftmp4]"=&f"(ftmp[4]), [tmp0]"=&r"(tmp[0]), + RESTRICT_ASM_ALL64 [block]"+&r"(block), [src1]"+&r"(src1), [src2]"+&r"(src2) : [stride]"r"((mips_reg)stride)