From patchwork Mon Sep 18 12:17:02 2017 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Manojkumar Bhosale X-Patchwork-Id: 5184 Delivered-To: ffmpegpatchwork@gmail.com Received: by 10.2.36.26 with SMTP id f26csp3382753jaa; Mon, 18 Sep 2017 05:17:15 -0700 (PDT) X-Google-Smtp-Source: ADKCNb6iADlrW4RdyjRj0PjeUEIlos64qtsu7Qsem9x6lCw8cQFC12B72zQEsIdPt6Bt4VRH3WhS X-Received: by 10.223.155.194 with SMTP id e2mr31190815wrc.49.1505737035334; Mon, 18 Sep 2017 05:17:15 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1505737035; cv=none; d=google.com; s=arc-20160816; b=PH17lghZdu6ODdXBrCtJDIRIMhs/lhLN54jnRJMIQLZTXOv3krYkeYoli/gU++NMxZ W4ex1njrwfQXcVXC5n/WSFI8lNyMfiYEBudyQHlJ4cjvUgfNl+/jzLc6zfOu46BsI78l sPqaCHkWKXO7xnwZJJ6fk0lQGwt8ZDliIbMk2e50r4H790/TmPNWv0OcnJ0iBBdMhUm5 CSdH6j/YhXSAGngQjWAZgoMPsO2VZVcvx55FcB9iyNXHkzJj0DkNAuTug+1IgCaXmYfX ChCHRY9cTOMyRBTnRpAfcO0T1kIt5CHuNxmRIokjI1VyjsAzRSxIYUHSxZXxU3vsvVS0 Rrhg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:content-transfer-encoding:cc:reply-to :list-subscribe:list-help:list-post:list-archive:list-unsubscribe :list-id:precedence:subject:mime-version:content-language :accept-language:in-reply-to:references:message-id:date:thread-index :thread-topic:to:from:delivered-to:arc-authentication-results; bh=7GIAkLcsZOD+/S+x3qOhPOQYtPEzcEFJBybYxkK6Bm0=; b=AY+0VxQjMwRDzq1GhvB0AGsbMZuFY9TZzBdmF3lcYPBcKnSzixrb+qU78Fe94l+cRI pXHuM1HA7yIgnRqGkUMrlOjNyMh1IGO9L8fzCZ+NRtXg/KxxMk1FgIVJ7NkT7ufV42O5 gSIpt0xESuKXCF6tIIwwvU14dLntoP5X/JPD1PhPbf6rGF1Sxi/Jxu9V8eC8nF+WJV+v 9S6GzQcEDJfJkDXW2gxJXf3Hdma5uMJp3rcWksbIoOH4YGHqF+EcNXP0ynIo7mQcZo43 7iVrvkSBM5Ee3yX8oRRWc7s5pMhYMdpMs9g23IV7ecSLDdi8oW9OP3MbhlP4HHB+pU3z EX2w== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org Return-Path: Received: from ffbox0-bg.mplayerhq.hu (ffbox0-bg.ffmpeg.org. [79.124.17.100]) by mx.google.com with ESMTP id b98si5826713wrd.541.2017.09.18.05.17.14; Mon, 18 Sep 2017 05:17:15 -0700 (PDT) Received-SPF: pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) client-ip=79.124.17.100; Authentication-Results: mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org Received: from [127.0.1.1] (localhost [127.0.0.1]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 3DC9168920E; Mon, 18 Sep 2017 15:17:05 +0300 (EEST) X-Original-To: ffmpeg-devel@ffmpeg.org Delivered-To: ffmpeg-devel@ffmpeg.org Received: from mailapp01.imgtec.com (mailapp01.imgtec.com [195.59.15.196]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 81F386891E9 for ; Mon, 18 Sep 2017 15:16:59 +0300 (EEST) Received: from HHMAIL01.hh.imgtec.org (unknown [10.100.10.19]) by Forcepoint Email with ESMTPS id BCD3A7DB00BD0 for ; Mon, 18 Sep 2017 13:17:02 +0100 (IST) Received: from PUMAIL01.pu.imgtec.org (192.168.91.250) by HHMAIL01.hh.imgtec.org (10.100.10.19) with Microsoft SMTP Server (TLS) id 14.3.294.0; Mon, 18 Sep 2017 13:17:05 +0100 Received: from PUMAIL01.pu.imgtec.org ([::1]) by PUMAIL01.pu.imgtec.org ([::1]) with mapi id 14.03.0266.001; Mon, 18 Sep 2017 17:47:03 +0530 From: Manojkumar Bhosale To: FFmpeg development discussions and patches Thread-Topic: [FFmpeg-devel] [PATCH] avcodec/mips: preload data in hevc sao edge 0 degree filter msa functions Thread-Index: AQHTMFd3PnVAS0KwAU+Oc+mDHFb8mqK6j1zg Date: Mon, 18 Sep 2017 12:17:02 +0000 Message-ID: <70293ACCC3BA6A4E81FFCA024C7A86E1E0591E60@PUMAIL01.pu.imgtec.org> References: <1505723080-30741-1-git-send-email-kaustubh.raste@imgtec.com> In-Reply-To: <1505723080-30741-1-git-send-email-kaustubh.raste@imgtec.com> Accept-Language: en-US Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [192.168.91.86] MIME-Version: 1.0 Subject: Re: [FFmpeg-devel] [PATCH] avcodec/mips: preload data in hevc sao edge 0 degree filter msa functions X-BeenThere: ffmpeg-devel@ffmpeg.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: FFmpeg development discussions and patches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Reply-To: FFmpeg development discussions and patches Cc: Kaustubh Raste Errors-To: ffmpeg-devel-bounces@ffmpeg.org Sender: "ffmpeg-devel" LGTM -----Original Message----- From: ffmpeg-devel [mailto:ffmpeg-devel-bounces@ffmpeg.org] On Behalf Of kaustubh.raste@imgtec.com Sent: Monday, September 18, 2017 1:55 PM To: ffmpeg-devel@ffmpeg.org Cc: Kaustubh Raste Subject: [FFmpeg-devel] [PATCH] avcodec/mips: preload data in hevc sao edge 0 degree filter msa functions From: Kaustubh Raste Signed-off-by: Kaustubh Raste --- libavcodec/mips/hevc_lpf_sao_msa.c | 232 +++++++++++++++++++++--------------- 1 file changed, 138 insertions(+), 94 deletions(-) SW(dst_val1, dst); dst += dst_stride; } + + src_minus10 = (v16u8) __msa_pckev_d((v2i64) src_minus11, + (v2i64) src_minus10); + + src0 = (v16i8) __msa_sldi_b(zero, (v16i8) src_minus10, 1); + src_plus10 = (v16i8) __msa_sldi_b(zero, (v16i8) src_minus10, 2); + + cmp_minus10 = ((v16u8) src0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus10 = ((v16u8) src0 == (v16u8) src_plus10); + diff_minus11 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = ((v16u8) src_plus10 < (v16u8) src0); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus10); + + offset = (v16i8) diff_minus10 + (v16i8) diff_minus11 + 2; + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset, offset, + offset, offset); + + src0 = (v16i8) __msa_xori_b((v16u8) src0, 128); + dst0 = __msa_adds_s_b(src0, offset); + dst0 = (v16i8) __msa_xori_b((v16u8) dst0, 128); + + dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); + dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); + + SW(dst_val0, dst); + dst += dst_stride; + SW(dst_val1, dst); } static void hevc_sao_edge_filter_0degree_8width_msa(uint8_t *dst, @@ -1329,64 +1357,90 @@ static void hevc_sao_edge_filter_0degree_8width_msa(uint8_t *dst, int16_t *sao_offset_val, int32_t height) { - uint8_t *src_minus1; - int32_t h_cnt; uint64_t dst_val0, dst_val1; - v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 + }; v16u8 const1 = (v16u8) __msa_ldi_b(1); - v16i8 dst0, dst1; - v16i8 zero = { 0 }; - v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; - v16u8 src_minus10, src_minus11; - v16i8 src_zero0, src_plus10, src_zero1, src_plus11; - v8i16 sao_offset, src00, offset_mask0, src01, offset_mask1; + v16u8 cmp_minus10, diff_minus10, diff_minus11; + v16u8 src0, src1, dst0, src_minus10, src_minus11, src_plus10, src_plus11; + v16i8 offset, sao_offset = LD_SB(sao_offset_val); - sao_offset = LD_SH(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); + src -= 1; - for (h_cnt = (height >> 1); h_cnt--;) { - src_minus1 = src - 1; - LD_UB2(src_minus1, src_stride, src_minus10, src_minus11); + /* load in advance */ + LD_UB2(src, src_stride, src_minus10, src_minus11); - SLDI_B2_0_SB(src_minus10, src_minus11, src_zero0, src_zero1, 1); - SLDI_B2_0_SB(src_minus10, src_minus11, src_plus10, src_plus11, 2); - ILVR_B2_UB(src_plus10, src_minus10, src_plus11, src_minus11, - src_minus10, src_minus11); - ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, - src_zero0, src_zero1); + for (height -= 2; height; height -= 2) { + src += (src_stride << 1); - cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + SLDI_B2_0_UB(src_minus10, src_minus11, src0, src1, 1); + SLDI_B2_0_UB(src_minus10, src_minus11, src_plus10, src_plus11, + 2); + + PCKEV_D2_UB(src_minus11, src_minus10, src_plus11, src_plus10, + src_minus10, src_plus10); + src0 = (v16u8) __msa_pckev_d((v2i64) src1, (v2i64) src0); + + cmp_minus10 = (src0 == src_minus10); diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); - cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + cmp_minus10 = (src_minus10 < src0); diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); - cmp_minus11 = ((v16u8) src_zero1 == src_minus11); - diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); - cmp_minus11 = (src_minus11 < (v16u8) src_zero1); - diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + cmp_minus10 = (src0 == src_plus10); + diff_minus11 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_plus10 < src0); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus10); - offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); - offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + offset = (v16i8) diff_minus10 + (v16i8) diff_minus11 + 2; - VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, - offset_mask0, offset_mask0, offset_mask0); - VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, - offset_mask1, offset_mask1, offset_mask1); - ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); + /* load in advance */ + LD_UB2(src, src_stride, src_minus10, src_minus11); - ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, - offset_mask1); - CLIP_SH2_0_255(offset_mask0, offset_mask1); - PCKEV_B2_SB(offset_mask0, offset_mask0, offset_mask1, offset_mask1, - dst0, dst1); + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset, offset, + offset, offset); + + src0 = __msa_xori_b(src0, 128); + dst0 = (v16u8) __msa_adds_s_b((v16i8) src0, offset); + dst0 = __msa_xori_b(dst0, 128); dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); - dst_val1 = __msa_copy_u_d((v2i64) dst1, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst0, 1); SD(dst_val0, dst); dst += dst_stride; SD(dst_val1, dst); dst += dst_stride; - src += (src_stride << 1); } + + SLDI_B2_0_UB(src_minus10, src_minus11, src0, src1, 1); + SLDI_B2_0_UB(src_minus10, src_minus11, src_plus10, src_plus11, 2); + + PCKEV_D2_UB(src_minus11, src_minus10, src_plus11, src_plus10, src_minus10, + src_plus10); + src0 = (v16u8) __msa_pckev_d((v2i64) src1, (v2i64) src0); + + cmp_minus10 = ((v16u8) src0 == src_minus10); + diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_minus10 < (v16u8) src0); + diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); + + cmp_minus10 = (src0 == src_plus10); + diff_minus11 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = (src_plus10 < src0); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus10); + + offset = (v16i8) diff_minus10 + (v16i8) diff_minus11 + 2; + + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset, offset, + offset, offset); + + src0 = __msa_xori_b(src0, 128); + dst0 = (v16u8) __msa_adds_s_b((v16i8) src0, offset); + dst0 = __msa_xori_b(dst0, 128); + + dst_val0 = __msa_copy_u_d((v2i64) dst0, 0); + dst_val1 = __msa_copy_u_d((v2i64) dst0, 1); + SD(dst_val0, dst); + dst += dst_stride; + SD(dst_val1, dst); } static void hevc_sao_edge_filter_0degree_16multiple_msa(uint8_t *dst, @@ -1398,7 +1452,7 @@ static void hevc_sao_edge_filter_0degree_16multiple_msa(uint8_t *dst, int32_t height) { uint8_t *dst_ptr, *src_minus1; - int32_t h_cnt, v_cnt; + int32_t v_cnt; v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; v16u8 const1 = (v16u8) __msa_ldi_b(1); v16i8 sao_offset; @@ -1411,20 +1465,18 @@ static void hevc_sao_edge_filter_0degree_16multiple_msa(uint8_t *dst, v16i8 offset_mask0, offset_mask1, offset_mask2, offset_mask3; v16i8 src_zero0, src_zero1, src_zero2, src_zero3; v16i8 src_plus10, src_plus11, src_plus12, src_plus13; - v8i16 src0, src1, src2, src3, src4, src5, src6, src7; - v8i16 temp0, temp1, temp2, temp3, temp4, temp5, temp6, temp7; sao_offset = LD_SB(sao_offset_val); sao_offset = __msa_pckev_b(sao_offset, sao_offset); - for (h_cnt = (height >> 2); h_cnt--;) { + for (; height; height -= 4) { src_minus1 = src - 1; LD_UB4(src_minus1, src_stride, src_minus10, src_minus11, src_minus12, src_minus13); - for (v_cnt = 0; v_cnt < (width >> 4); v_cnt++) { + for (v_cnt = 0; v_cnt < width; v_cnt += 16) { src_minus1 += 16; - dst_ptr = dst + (v_cnt << 4); + dst_ptr = dst + v_cnt; LD_UB4(src_minus1, src_stride, src10, src11, src12, src13); SLDI_B2_SB(src10, src11, src_minus10, src_minus11, src_zero0, @@ -1485,22 +1537,14 @@ static void hevc_sao_edge_filter_0degree_16multiple_msa(uint8_t *dst, VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask3, offset_mask3, offset_mask3, offset_mask3); - UNPCK_UB_SH(src_zero0, src0, src1); - UNPCK_SB_SH(offset_mask0, temp0, temp1); - UNPCK_UB_SH(src_zero1, src2, src3); - UNPCK_SB_SH(offset_mask1, temp2, temp3); - UNPCK_UB_SH(src_zero2, src4, src5); - UNPCK_SB_SH(offset_mask2, temp4, temp5); - UNPCK_UB_SH(src_zero3, src6, src7); - UNPCK_SB_SH(offset_mask3, temp6, temp7); - ADD4(temp0, src0, temp1, src1, temp2, src2, temp3, src3, temp0, - temp1, temp2, temp3); - ADD4(temp4, src4, temp5, src5, temp6, src6, temp7, src7, temp4, - temp5, temp6, temp7); - CLIP_SH4_0_255(temp0, temp1, temp2, temp3); - CLIP_SH4_0_255(temp4, temp5, temp6, temp7); - PCKEV_B4_UB(temp1, temp0, temp3, temp2, temp5, temp4, temp7, temp6, - dst0, dst1, dst2, dst3); + XORI_B4_128_SB(src_zero0, src_zero1, src_zero2, src_zero3); + + dst0 = (v16u8) __msa_adds_s_b((v16i8) src_zero0, offset_mask0); + dst1 = (v16u8) __msa_adds_s_b((v16i8) src_zero1, offset_mask1); + dst2 = (v16u8) __msa_adds_s_b((v16i8) src_zero2, offset_mask2); + dst3 = (v16u8) __msa_adds_s_b((v16i8) src_zero3, + offset_mask3); + + XORI_B4_128_UB(dst0, dst1, dst2, dst3); src_minus10 = src10; ST_UB(dst0, dst_ptr); -- 1.7.9.5 _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org http://ffmpeg.org/mailman/listinfo/ffmpeg-devel diff --git a/libavcodec/mips/hevc_lpf_sao_msa.c b/libavcodec/mips/hevc_lpf_sao_msa.c index 1d77432..3472d32 100644 --- a/libavcodec/mips/hevc_lpf_sao_msa.c +++ b/libavcodec/mips/hevc_lpf_sao_msa.c @@ -1265,54 +1265,51 @@ static void hevc_sao_edge_filter_0degree_4width_msa(uint8_t *dst, int16_t *sao_offset_val, int32_t height) { - int32_t h_cnt; uint32_t dst_val0, dst_val1; - v8i16 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0 }; + v16u8 cmp_minus10, diff_minus10, diff_minus11, src_minus10, src_minus11; + v16i8 edge_idx = { 1, 2, 0, 3, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; + v16i8 sao_offset = LD_SB(sao_offset_val); + v16i8 src_plus10, offset, src0, dst0; v16u8 const1 = (v16u8) __msa_ldi_b(1); - v16u8 cmp_minus10, diff_minus10, cmp_minus11, diff_minus11; - v16u8 src_minus10, src_minus11; v16i8 zero = { 0 }; - v16i8 src_zero0, src_zero1, src_plus10, src_plus11, dst0; - v8i16 offset_mask0, offset_mask1; - v8i16 sao_offset, src00, src01; - sao_offset = LD_SH(sao_offset_val); + sao_offset = __msa_pckev_b(sao_offset, sao_offset); src -= 1; - for (h_cnt = (height >> 1); h_cnt--;) { - LD_UB2(src, src_stride, src_minus10, src_minus11); + /* load in advance */ + LD_UB2(src, src_stride, src_minus10, src_minus11); + + for (height -= 2; height; height -= 2) { src += (2 * src_stride); - SLDI_B2_0_SB(src_minus10, src_minus11, src_zero0, src_zero1, 1); - SLDI_B2_0_SB(src_minus10, src_minus11, src_plus10, src_plus11, 2); - ILVR_B2_UB(src_plus10, src_minus10, src_plus11, src_minus11, - src_minus10, src_minus11); - ILVR_B2_SB(src_zero0, src_zero0, src_zero1, src_zero1, src_zero0, - src_zero1); + src_minus10 = (v16u8) __msa_pckev_d((v2i64) src_minus11, + (v2i64) src_minus10); - cmp_minus10 = ((v16u8) src_zero0 == src_minus10); + src0 = (v16i8) __msa_sldi_b(zero, (v16i8) src_minus10, 1); + src_plus10 = (v16i8) __msa_sldi_b(zero, (v16i8) src_minus10, + 2); + + cmp_minus10 = ((v16u8) src0 == src_minus10); diff_minus10 = __msa_nor_v(cmp_minus10, cmp_minus10); - cmp_minus10 = (src_minus10 < (v16u8) src_zero0); + cmp_minus10 = (src_minus10 < (v16u8) src0); diff_minus10 = __msa_bmnz_v(diff_minus10, const1, cmp_minus10); - cmp_minus11 = ((v16u8) src_zero1 == src_minus11); - diff_minus11 = __msa_nor_v(cmp_minus11, cmp_minus11); - cmp_minus11 = (src_minus11 < (v16u8) src_zero1); - diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus11); + cmp_minus10 = ((v16u8) src0 == (v16u8) src_plus10); + diff_minus11 = __msa_nor_v(cmp_minus10, cmp_minus10); + cmp_minus10 = ((v16u8) src_plus10 < (v16u8) src0); + diff_minus11 = __msa_bmnz_v(diff_minus11, const1, cmp_minus10); - offset_mask0 = (v8i16) (__msa_hadd_u_h(diff_minus10, diff_minus10) + 2); - offset_mask1 = (v8i16) (__msa_hadd_u_h(diff_minus11, diff_minus11) + 2); + offset = (v16i8) diff_minus10 + (v16i8) diff_minus11 + 2; - VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask0, - offset_mask0, offset_mask0, offset_mask0); - VSHF_H2_SH(edge_idx, edge_idx, sao_offset, sao_offset, offset_mask1, - offset_mask1, offset_mask1, offset_mask1); - ILVEV_B2_SH(src_zero0, zero, src_zero1, zero, src00, src01); - ADD2(offset_mask0, src00, offset_mask1, src01, offset_mask0, - offset_mask1); - CLIP_SH2_0_255(offset_mask0, offset_mask1); + /* load in advance */ + LD_UB2(src, src_stride, src_minus10, src_minus11); + + VSHF_B2_SB(edge_idx, edge_idx, sao_offset, sao_offset, offset, offset, + offset, offset); + + src0 = (v16i8) __msa_xori_b((v16u8) src0, 128); + dst0 = __msa_adds_s_b(src0, offset); + dst0 = (v16i8) __msa_xori_b((v16u8) dst0, 128); - dst0 = __msa_pckev_b((v16i8) offset_mask1, (v16i8) offset_mask0); dst_val0 = __msa_copy_u_w((v4i32) dst0, 0); dst_val1 = __msa_copy_u_w((v4i32) dst0, 2); SW(dst_val0, dst); @@ -1320,6 +1317,37 @@ static void hevc_sao_edge_filter_0degree_4width_msa(uint8_t *dst,