From patchwork Wed Aug 4 02:06:15 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Wu Jianhua X-Patchwork-Id: 29228 Delivered-To: ffmpegpatchwork2@gmail.com Received: by 2002:a6b:6c0f:0:0:0:0:0 with SMTP id a15csp2976191ioh; Tue, 3 Aug 2021 19:07:06 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwwkgnP9DxdV+vlmEzYcY6yGoDqAD6Mmati+8tAU/KvOQsaHMaOmdAA2FDUTJ2EXZWLLm5s X-Received: by 2002:a50:ab1c:: with SMTP id s28mr28372491edc.214.1628042825942; Tue, 03 Aug 2021 19:07:05 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1628042825; cv=none; d=google.com; s=arc-20160816; b=vc+NEuaQDl5ffqXPXQxQ1OBQ5obTAa/diAHNS6uuzTQ8N1jlSBX1Cg3hXrfo4bjEdk 3iREN2iT+vOssIy7mgqbC+Bl8g9jdhQAKixux2uoz2X2OvOGB8n8kBLkoYabux5UlsFb VRlEY+C4hljgkqNN8MZ04MgPUN4tAsQd6CRq3yDt8130PNXR+BknqDRpe+I1UGty8rQq l6Krp5YIncpzir7ELfp+RTdREcxtM+lrDpQsAbPKgD1Qp21qICui1GN4srSGG3OnFtHi leXiVEzRQ6OV46Wrbu7GPgPwD3RBEniq4MDyjChOU47QOtopcB3NCj4pbk+2jTUxIrad 325Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:content-transfer-encoding:mime-version:cc:reply-to :list-subscribe:list-help:list-post:list-archive:list-unsubscribe :list-id:precedence:subject:references:in-reply-to:message-id:date :to:from:delivered-to; bh=PboxEQarOcXBkRjYc7ME5/yZRELmlW30wtsi4MykEFQ=; b=wmsKlQePyDiRxd9LotMlzuOQzx0oviSGi1ilYno8KeuY+QFcYGW+/biWaaVyH8m87x NCevFHMz2YeodX+AfUe2J6jjgsJ0i8EJyeX96i+KCYOtCIJnGcQachOIiPkm7AA5smU3 ckahRmEk5Rk4aHjt4bEiurfg/EzqV1s+PvGQh3YrPe7B4/xGRnDSSq3zK1Pd6VoLuFG7 0NT+iVB/msk7fBXY4LyJgvE2tLw6T/Hz94PaZud8kf0NW7HYS1qe61bFfmyb8iKry7H6 r3uJIpcPXQWxpZvVdkVmAAbIuiZuMt4XiDD4jC3WzoUQPkcFBqdShHti2r+rpEEcpm0c IYag== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from ffbox0-bg.mplayerhq.hu (ffbox0-bg.ffmpeg.org. [79.124.17.100]) by mx.google.com with ESMTP id kl11si679971ejc.593.2021.08.03.19.07.05; Tue, 03 Aug 2021 19:07:05 -0700 (PDT) Received-SPF: pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) client-ip=79.124.17.100; Authentication-Results: mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: from [127.0.1.1] (localhost [127.0.0.1]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 03FB3680371; Wed, 4 Aug 2021 05:06:39 +0300 (EEST) X-Original-To: ffmpeg-devel@ffmpeg.org Delivered-To: ffmpeg-devel@ffmpeg.org Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTPS id B1B6E689955 for ; Wed, 4 Aug 2021 05:06:31 +0300 (EEST) X-IronPort-AV: E=McAfee;i="6200,9189,10065"; a="210717028" X-IronPort-AV: E=Sophos;i="5.84,293,1620716400"; d="scan'208";a="210717028" Received: from fmsmga006.fm.intel.com ([10.253.24.20]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 03 Aug 2021 19:06:26 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.84,293,1620716400"; d="scan'208";a="667632145" Received: from skl-e5.sh.intel.com ([10.239.43.106]) by fmsmga006.fm.intel.com with ESMTP; 03 Aug 2021 19:06:25 -0700 From: Wu Jianhua To: ffmpeg-devel@ffmpeg.org Date: Wed, 4 Aug 2021 10:06:15 +0800 Message-Id: <20210804020616.82866-4-jianhua.wu@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20210804020616.82866-1-jianhua.wu@intel.com> References: <20210804020616.82866-1-jianhua.wu@intel.com> Subject: [FFmpeg-devel] [PATCH v2 4/5] libavfilter/x86/vf_gblur: add localbuf and ff_horiz_slice_avx2/512() X-BeenThere: ffmpeg-devel@ffmpeg.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: FFmpeg development discussions and patches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Reply-To: FFmpeg development discussions and patches Cc: Wu Jianhua MIME-Version: 1.0 Errors-To: ffmpeg-devel-bounces@ffmpeg.org Sender: "ffmpeg-devel" X-TUID: xqrPgiOcweLG We introduced a ff_horiz_slice_avx2/512() implemented on a new algorithm. In a nutshell, the new algorithm does three things, gathering data from 8/16 rows, blurring data, and scattering data back to the image buffer. Here we used a customized transpose 8x8/16x16 to avoid the huge overhead brought by gather and scatter instructions, which is dependent on the temporary buffer called localbuf added newly. Performance data: ff_horiz_slice_avx2(old): 109.89 ff_horiz_slice_avx2(new): 666.67 ff_horiz_slice_avx512: 1000 Co-authored-by: Cheng Yanfei Co-authored-by: Jin Jun Signed-off-by: Wu Jianhua --- libavfilter/gblur.h | 4 +- libavfilter/vf_gblur.c | 11 +- libavfilter/x86/vf_gblur.asm | 579 +++++++++++++++++++++++++++++++- libavfilter/x86/vf_gblur_init.c | 17 +- 4 files changed, 601 insertions(+), 10 deletions(-) diff --git a/libavfilter/gblur.h b/libavfilter/gblur.h index 367575a6db..3a66984b06 100644 --- a/libavfilter/gblur.h +++ b/libavfilter/gblur.h @@ -39,9 +39,11 @@ typedef struct GBlurContext { int flt; int depth; + int stride; int planewidth[4]; int planeheight[4]; float *buffer; + float *localbuf; ///< temporary buffer for horiz_slice. NULL if not used float boundaryscale; float boundaryscaleV; float postscale; @@ -49,7 +51,7 @@ typedef struct GBlurContext { float nu; float nuV; int nb_planes; - void (*horiz_slice)(float *buffer, int width, int height, int steps, float nu, float bscale); + void (*horiz_slice)(float *buffer, int width, int height, int steps, float nu, float bscale, float *localbuf); void (*verti_slice)(float *buffer, int width, int height, int slice_start, int slice_end, int steps, float nu, float bscale); void (*postscale_slice)(float *buffer, int length, float postscale, float min, float max); diff --git a/libavfilter/vf_gblur.c b/libavfilter/vf_gblur.c index de7ed82d49..0768fe12e1 100644 --- a/libavfilter/vf_gblur.c +++ b/libavfilter/vf_gblur.c @@ -64,7 +64,7 @@ static void postscale_c(float *buffer, int length, } static void horiz_slice_c(float *buffer, int width, int height, int steps, - float nu, float bscale) + float nu, float bscale, float *localbuf) { int step, x, y; float *ptr; @@ -97,9 +97,13 @@ static int filter_horizontally(AVFilterContext *ctx, void *arg, int jobnr, int n const int steps = s->steps; const float nu = s->nu; float *buffer = s->buffer; + float *localbuf = NULL; + + if (s->localbuf) + localbuf = s->localbuf + s->stride * width * slice_start; s->horiz_slice(buffer + width * slice_start, width, slice_end - slice_start, - steps, nu, boundaryscale); + steps, nu, boundaryscale, localbuf); emms_c(); return 0; } @@ -239,6 +243,7 @@ static int query_formats(AVFilterContext *ctx) void ff_gblur_init(GBlurContext *s) { + s->localbuf = NULL; s->horiz_slice = horiz_slice_c; s->verti_slice = verti_slice_c; s->postscale_slice = postscale_c; @@ -381,6 +386,8 @@ static av_cold void uninit(AVFilterContext *ctx) GBlurContext *s = ctx->priv; av_freep(&s->buffer); + if (s->localbuf) + av_free(s->localbuf); } static const AVFilterPad gblur_inputs[] = { diff --git a/libavfilter/x86/vf_gblur.asm b/libavfilter/x86/vf_gblur.asm index ac4debba74..c0d57cc82b 100644 --- a/libavfilter/x86/vf_gblur.asm +++ b/libavfilter/x86/vf_gblur.asm @@ -20,6 +20,14 @@ %include "libavutil/x86/x86util.asm" +SECTION .data + +gblur_transpose_16x16_indices1: dq 2, 3, 0, 1, 6, 7, 4, 5 +gblur_transpose_16x16_indices2: dq 1, 0, 3, 2, 5, 4, 7, 6 +gblur_transpose_16x16_indices3: dd 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 +gblur_transpose_16x16_mask: dw 0xcc, 0x33, 0xaa, 0x55, 0xaaaa, 0x5555 +gblur_vindex_width: dd 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 + SECTION .text %xdefine AVX2_MMSIZE 32 @@ -32,6 +40,29 @@ SECTION .text %endrep %endmacro +%macro KXNOR 2-* +%if mmsize == AVX512_MMSIZE + kxnorw %2, %2, %2 +%else + %if %0 == 3 + mov %3, -1 + %else + vpcmpeqd %1, %1, %1 + %endif +%endif +%endmacro + +%macro KMOVW 2-4 +%if mmsize == AVX2_MMSIZE && %0 == 4 + mova %1, %2 +%elif mmsize == AVX512_MMSIZE + %if %0 == 4 + %rotate 2 + %endif + kmovw %1, %2 +%endif +%endmacro + %macro PUSH_MASK 5 %if mmsize == AVX2_MMSIZE %assign %%n mmsize/4 @@ -59,15 +90,546 @@ SECTION .text %endif %endmacro -; void ff_horiz_slice_sse4(float *ptr, int width, int height, int steps, -; float nu, float bscale) +%macro VGATHERDPS 4 +%if mmsize == AVX2_MMSIZE + vgatherdps %1, %2, %3 +%else + vgatherdps %1{%4}, %2 +%endif +%endmacro + +%macro VSCATTERDPS128 7 + %rep 4 + mov %7, %6 + and %7, 1 + cmp %7, 0 + je %%end_scatter + movss [%2 + %3*%4], xm%1 + vpshufd m%1, m%1, 0x39 + add %3, %5 + sar %6, 1 + %endrep + %%end_scatter: +%endmacro + +; %1=register index +; %2=base address %3=vindex +; %4=scale %5=width +; %6=mask %7=tmp +; m15=reserved +%macro VSCATTERDPS256 7 + mova m15, m%1 + xor %3, %3 + VSCATTERDPS128 15, %2, %3, %4, %5, %6, %7 + vextractf128 xm15, m%1, 1 + VSCATTERDPS128 15, %2, %3, %4, %5, %6, %7 +%endmacro + +; %1=base address %2=avx2 vindex +; %3=avx512 vindex %4=avx2 mask +; %5=avx512 mask %6=register index +; %7=width %8-*=tmp +%macro VSCATTERDPS 8-* +%if mmsize == AVX2_MMSIZE + %if %0 == 9 + mov %9, %4 + VSCATTERDPS256 %6, %1, %2, 4, %7, %9, %8 + %else + VSCATTERDPS256 %6, %1, %2, 4, %7, %4, %8 + %endif +%else + vscatterdps [%1 + %3*4]{%5}, m%6 +%endif +%endmacro + +%macro INIT_WORD_MASK 1-* + %assign %%i 0 + %rep %0 + kmovw %1, [gblur_transpose_16x16_mask + %%i * 2] + %assign %%i %%i+1 + %rotate 1 + %endrep +%endmacro + +%macro INIT_INDICES 1-* + %assign %%i 1 + %rep %0 + movu %1, [gblur_transpose_16x16_indices %+ %%i] + %assign %%i %%i+1 + %rotate 1 + %endrep +%endmacro + +%assign stack_offset 0 +%macro PUSH_MM 1 +%if mmsize == AVX2_MMSIZE + movu [rsp + stack_offset], %1 + %assign stack_offset stack_offset+mmsize +%endif +%endmacro + +%macro POP_MM 1 +%if mmsize == AVX2_MMSIZE + %assign stack_offset stack_offset-mmsize + movu %1, [rsp + stack_offset] +%endif +%endmacro + +%macro READ_LOCAL_BUFFER 1 + %if mmsize == AVX512_MMSIZE + %assign %%i 19 + %else + %assign %%i 9 + %endif + %assign %%j %%i-1 + %assign %%k %1-1 + %xdefine %%m m %+ %%i + mova %%m, m3 + FMULADD_PS %%m, %%m, m0, [localbufq + %%k * mmsize], %%m + %assign %%k %%k-1 + %rep %1-1 + %xdefine %%m m %+ %%j + mova %%m, m %+ %%i + FMULADD_PS %%m, %%m, m0, [localbufq + %%k * mmsize], %%m + %assign %%i %%i-1 + %assign %%j %%j-1 + %assign %%k %%k-1 + %endrep + %if mmsize == AVX512_MMSIZE + mova m3, m %+ %%i + %endif +%endmacro + +%macro FMADD_WRITE 4 + FMULADD_PS %1, %1, %2, %3, %1 + mova %4, %1 +%endmacro + +%macro WRITE_LOCAL_BUFFER_INTERNAL 8-16 + %assign %%i 0 + %rep %0 + FMADD_WRITE m3, m0, m %+ %1, [localbufq + %%i * mmsize] + %assign %%i %%i+1 + %rotate 1 + %endrep +%endmacro + +%macro GATHERPS 1 + %if mmsize == AVX512_MMSIZE + %assign %%i 4 + %else + %assign %%i 2 + %endif + movu m %+ %%i, [ptrq] + mov strideq, widthq + %assign %%i %%i+1 + %rep %1-2 + movu m %+ %%i, [ptrq + strideq*4] + add strideq, widthq + %assign %%i %%i+1 + %endrep + movu m %+ %%i, [ptrq + strideq*4] +%endmacro + +%macro SCATTERPS_INTERNAL 8-16 + movu [ptrq + strideq*0], m %+ %1 + mov strideq, widthq + %rotate 1 + %rep %0-2 + movu [ptrq + strideq*4], m %+ %1 + add strideq, widthq + %rotate 1 + %endrep + movu [ptrq + strideq*4], m %+ %1 +%endmacro + +%macro BATCH_INSERT64X4 4-* + %assign %%imm8 %1 + %rotate 1 + %rep (%0-1)/3 + vinserti64x4 m%1, m%2, ym%3, %%imm8 + %rotate 3 + %endrep +%endmacro + +%macro BATCH_EXTRACT_INSERT 2-* + %assign %%imm8 %1 + %rotate 1 + %rep (%0-1)/2 + vextractf64x4 ym%1, m%1, %%imm8 + vextractf64x4 ym%2, m%2, %%imm8 + vinserti64x4 m%1, m%1, ym%2, %%imm8 + %rotate 2 + %endrep +%endmacro + +%macro BATCH_MOVE 2-* + %rep %0/2 + mova m%1, m%2 + %rotate 2 + %endrep +%endmacro + +%macro BATCH_PERMUTE 3-* + %xdefine %%decorator %1 + %xdefine %%mask %2 + %assign %%index %3 + %rotate 3 + %rep (%0-3)/2 + vperm %+ %%decorator m%1{%%mask}, m %+ %%index, m%2 + %rotate 2 + %endrep +%endmacro +; input : m3-m19 +; output: m8 m5 m9 m15 m16 m7 m17 m27 m24 m21 m25 m19 m12 m23 m13 m11 +%macro TRANSPOSE_16X16_AVX512 0 + BATCH_INSERT64X4 0x1, 20,4,12, 21,5,13, 22,6,14, 23,7,15 + BATCH_INSERT64X4 0x1, 24,8,16, 25,9,17, 26,10,18, 27,11,19 + + BATCH_EXTRACT_INSERT 0x1, 4,12, 5,13, 6,14, 7,15 + BATCH_EXTRACT_INSERT 0x1, 8,16, 9,17, 10,18, 11,19 + + BATCH_MOVE 12,20, 13,21, 14,22, 15,23 + BATCH_PERMUTE q, k6, 28, 12,24, 13,25, 14,26, 15,27 + BATCH_PERMUTE q, k5, 28, 24,20, 25,21, 26,22, 27,23 + + BATCH_MOVE 16,4, 17,5, 18,6, 19,7 + BATCH_PERMUTE q, k6, 28, 16,8, 17,9, 18,10, 19,11 + BATCH_PERMUTE q, k5, 28, 8,4, 9,5, 10,6, 11,7 + + BATCH_MOVE 4,12, 5,13, 6,24, 7,25 + BATCH_MOVE 20,16, 21,17, 22,8, 23,9 + + BATCH_PERMUTE q, k4, 29, 4,14, 5,15, 6,26, 7,27 + BATCH_PERMUTE q, k3, 29, 14,12, 15,13, 26,24, 27,25 + BATCH_PERMUTE q, k4, 29, 20,18, 21,19, 22,10, 23,11 + BATCH_PERMUTE q, k3, 29, 18,16, 19,17, 10,8, 11,9 + + BATCH_MOVE 8,4, 9,14, 16,6, 17,26 + BATCH_MOVE 24,20, 25,18, 12,22, 13,10 + + BATCH_PERMUTE d, k2, 30, 8,5, 9,15, 16,7, 17,27 + BATCH_PERMUTE d, k1, 30, 5,4, 15,14, 7,6, 27,26 + BATCH_PERMUTE d, k2, 30, 24,21, 25,19, 12,23, 13,11 + BATCH_PERMUTE d, k1, 30, 21,20, 19,18, 23,22, 11,10 +%endmacro + +%macro INSERT_UNPACK 8 + vinsertf128 m%5, m%1, xm%3, 0x1 + vinsertf128 m%6, m%2, xm%4, 0x1 + vunpcklpd m%7, m%5, m%6 + vunpckhpd m%8, m%5, m%6 +%endmacro + +%macro SHUFFLE 4 + vshufps m%3, m%1, m%2, 0x88 + vshufps m%4, m%1, m%2, 0xDD + mova m%1, m%3 + mova m%2, m%4 +%endmacro + +%macro EXTRACT_INSERT_UNPACK 6 + vextractf128 xm%1, m%1, 0x1 + vextractf128 xm%2, m%2, 0x1 + vinsertf128 m%3, m%3, xm%1, 0x0 + vinsertf128 m%4, m%4, xm%2, 0x0 + vunpcklpd m%5, m%3, m%4 + vunpckhpd m%6, m%3, m%4 +%endmacro + +; Transpose 8x8 AVX2 +; Limit the number ym# register to 16 for compatibility +; Used up registers instead of using stack memory +; Input: m2-m9 +; Output: m12, m14, m13, m15, m8, m10, m9, m11 +%macro TRANSPOSE_8X8_AVX2 0 + INSERT_UNPACK 2, 3, 6, 7, 10, 11, 12, 13 + INSERT_UNPACK 4, 5, 8, 9, 10, 11, 14, 15 + + SHUFFLE 12, 14, 10, 11 + SHUFFLE 13, 15, 10, 11 + + EXTRACT_INSERT_UNPACK 4, 5, 8, 9, 10, 11 + EXTRACT_INSERT_UNPACK 2, 3, 6, 7, 8, 9 + + SHUFFLE 8, 10, 6, 7 + SHUFFLE 9, 11, 6, 7 +%endmacro + +%macro TRANSPOSE 0 + %if cpuflag(avx512) + TRANSPOSE_16X16_AVX512 + %elif cpuflag(avx2) + TRANSPOSE_8X8_AVX2 + %endif +%endmacro + +%macro WRITE_LOCAL_BUFFER 0 + %if cpuflag(avx512) + WRITE_LOCAL_BUFFER_INTERNAL 8, 5, 9, 15, 16, 7, 17, 27, \ + 24, 21, 25, 19, 12, 23, 13, 11 + %elif cpuflag(avx2) + WRITE_LOCAL_BUFFER_INTERNAL 12, 14, 13, 15, 8, 10, 9, 11 + %endif +%endmacro + +%macro SCATTERPS 0 + %if cpuflag(avx512) + SCATTERPS_INTERNAL 8, 5, 9, 15, 16, 7, 17, 27, \ + 24, 21, 25, 19, 12, 23, 13, 11 + %elif cpuflag(avx2) + SCATTERPS_INTERNAL 12, 14, 13, 15, 8, 10, 9, 11 + %endif +%endmacro + +%macro OPTIMIZED_LOOP_STEP 0 + lea stepd, [stepsd - 1] + cmp stepd, 0 + jle %%bscale_scalar +%%loop_step: + sub localbufq, mmsize + mulps m3, m1 + movu [localbufq], m3 + + ; Filter leftwards + lea xq, [widthq - 1] + %%loop_step_x_back: + sub localbufq, mmsize + FMULADD_PS m3, m3, m0, [localbufq], m3 + movu [localbufq], m3 + + dec xq + cmp xq, 0 + jg %%loop_step_x_back + + ; Filter rightwards + mulps m3, m1 + movu [localbufq], m3 + add localbufq, mmsize + + lea xq, [widthq - 1] + %%loop_step_x: + FMULADD_PS m3, m3, m0, [localbufq], m3 + movu [localbufq], m3 + add localbufq, mmsize + + dec xq + cmp xq, 0 + jg %%loop_step_x + + dec stepd + cmp stepd, 0 + jg %%loop_step + +%%bscale_scalar: +%endmacro + +;*************************************************************************** +; void ff_horiz_slice(float *ptr, int width, int height, int steps, +; float nu, float bscale) +;*************************************************************************** %macro HORIZ_SLICE 0 %if UNIX64 +%if cpuflag(avx512) || cpuflag(avx2) +cglobal horiz_slice, 5, 12, mmnum, 0-mmsize*4, buffer, width, height, steps, \ + localbuf, x, y, step, stride, remain, ptr, mask +%else cglobal horiz_slice, 4, 9, 9, ptr, width, height, steps, x, y, step, stride, remain +%endif +%else +%if cpuflag(avx512) || cpuflag(avx2) +cglobal horiz_slice, 5, 12, mmnum, 0-mmsize*4, buffer, width, height, steps, nu, bscale, \ + localbuf, x, y, step, stride, remain, ptr, mask %else cglobal horiz_slice, 4, 9, 9, ptr, width, height, steps, nu, bscale, x, y, step, stride, remain %endif +%endif +%if cpuflag(avx512) || cpuflag(avx2) +%assign rows mmsize/4 +%assign cols mmsize/4 +%if WIN64 + VBROADCASTSS m0, num ; nu + VBROADCASTSS m1, bscalem ; bscale + + mov nuq, localbufm + DEFINE_ARGS buffer, width, height, steps, \ + localbuf, x, y, step, stride, remain, ptr, mask + MOVSXDIFNIDN width, height, steps +%else + VBROADCASTSS m0, xmm0 ; nu + VBROADCASTSS m1, xmm1 ; bscale +%endif + +%if cpuflag(avx512) + vpbroadcastd m2, widthd + INIT_WORD_MASK k6, k5, k4, k3, k2, k1 + INIT_INDICES m28, m29, m30 +%else + movd xm2, widthd + VBROADCASTSS m2, xm2 +%endif + + vpmulld m2, m2, [gblur_vindex_width] ; vindex width + + xor yq, yq ; y = 0 + xor xq, xq ; x = 0 + + cmp heightq, rows + jl .y_scalar + sub heightq, rows + +.loop_y: + ; ptr = buffer + y * width; + mov ptrq, yq + imul ptrq, widthq + lea ptrq, [bufferq + ptrq*4] + + KXNOR m5, k7 + VGATHERDPS m3, [ptrq + m2*4], m5, k7 + mulps m3, m1 + movu [localbufq], m3 + add ptrq, 4 + add localbufq, mmsize + + ; Filter rightwards + PUSH_MM m2 + lea xq, [widthq - 1] + .loop_x: + PUSH_MM m3 + GATHERPS cols + TRANSPOSE + POP_MM m3 + WRITE_LOCAL_BUFFER + + add ptrq, mmsize + add localbufq, rows * mmsize + sub xq, cols + cmp xq, cols + jge .loop_x + POP_MM m2 + + cmp xq, 0 + jle .bscale_scalar + .loop_x_scalar: + KXNOR m5, k7 + VGATHERDPS m4, [ptrq + m2*4], m5, k7 + FMULADD_PS m3, m3, m0, m4, m3 + movu [localbufq], m3 + + add ptrq, 0x4 + add localbufq, mmsize + dec xq + cmp xq, 0 + jg .loop_x_scalar + + OPTIMIZED_LOOP_STEP + + .bscale_scalar: + sub ptrq, 4 + sub localbufq, mmsize + mulps m3, m1 + KXNOR m5, k7, maskq + VSCATTERDPS ptrq, strideq, m2, maskq, k7, 3, widthq, remainq + + ; Filter leftwards + PUSH_MM m2 + lea xq, [widthq - 1] + .loop_x_back: + sub localbufq, rows * mmsize + READ_LOCAL_BUFFER cols + PUSH_MM m2 + TRANSPOSE + POP_MM m3 + sub ptrq, mmsize + SCATTERPS + + sub xq, cols + cmp xq, cols + jge .loop_x_back + POP_MM m2 + + cmp xq, 0 + jle .end_loop_x + .loop_x_back_scalar: + sub ptrq, 0x4 + sub localbufq, mmsize + FMULADD_PS m3, m3, m0, [localbufq], m3 + KXNOR m5, k7, maskq + VSCATTERDPS ptrq, strideq, m2, maskq, k7, 3, widthq, remainq + + dec xq + cmp xq, 0 + jg .loop_x_back_scalar + + .end_loop_x: + + add yq, rows + cmp yq, heightq + jle .loop_y + + add heightq, rows + cmp yq, heightq + jge .end_scalar + + mov remainq, widthq + imul remainq, mmsize + add ptrq, remainq + +.y_scalar: + mov remainq, heightq + sub remainq, yq + mov maskq, 1 + shlx maskq, maskq, remainq + sub maskq, 1 + mov remainq, maskq + PUSH_MASK m5, k1, remaind, xd, rsp + 0x20 + + mov ptrq, yq + imul ptrq, widthq + lea ptrq, [bufferq + ptrq * 4] ; ptrq = buffer + y * width + KMOVW m6, m5, k7, k1 + VGATHERDPS m3, [ptrq + m2 * 4], m6, k7 + mulps m3, m1 ; p0 *= bscale + movu [localbufq], m3 + add localbufq, mmsize + + ; Filter rightwards + lea xq, [widthq - 1] + .y_scalar_loop_x: + add ptrq, 4 + KMOVW m6, m5, k7, k1 + VGATHERDPS m4, [ptrq + m2 * 4], m6, k7 + FMULADD_PS m3, m3, m0, m4, m3 + movu [localbufq], m3 + add localbufq, mmsize + + dec xq + cmp xq, 0 + jg .y_scalar_loop_x + + OPTIMIZED_LOOP_STEP + + sub localbufq, mmsize + mulps m3, m1 ; p0 *= bscale + KMOVW k7, k1 + VSCATTERDPS ptrq, strideq, m2, maskq, k7, 3, widthq, remainq, heightq + + ; Filter leftwards + lea xq, [widthq - 1] + .y_scalar_loop_x_back: + sub ptrq, 4 + sub localbufq, mmsize + FMULADD_PS m3, m3, m0, [localbufq], m3 + KMOVW k7, k1 + VSCATTERDPS ptrq, strideq, m2, maskq, k7, 3, widthq, remainq, heightq + dec xq + cmp xq, 0 + jg .y_scalar_loop_x_back + +.end_scalar: + RET +%else %if WIN64 movss m0, num movss m1, bscalem @@ -211,16 +773,26 @@ cglobal horiz_slice, 4, 9, 9, ptr, width, height, steps, nu, bscale, x, y, step, jl .loop_y RET +%endif %endmacro %if ARCH_X86_64 INIT_XMM sse4 HORIZ_SLICE -INIT_XMM avx2 +%if HAVE_AVX2_EXTERNAL +INIT_YMM avx2 +%xdefine mmnum 16 HORIZ_SLICE %endif +%if HAVE_AVX512_EXTERNAL +INIT_ZMM avx512 +%xdefine mmnum 32 +HORIZ_SLICE +%endif +%endif + %macro POSTSCALE_SLICE 0 cglobal postscale_slice, 2, 2, 4, ptr, length, postscale, min, max shl lengthd, 2 @@ -270,7 +842,6 @@ INIT_ZMM avx512 POSTSCALE_SLICE %endif - ;******************************************************************************* ; void ff_verti_slice(float *buffer, int width, int height, int column_begin, ; int column_end, int steps, float nu, float bscale); diff --git a/libavfilter/x86/vf_gblur_init.c b/libavfilter/x86/vf_gblur_init.c index 3e173410c2..b47f6fbffb 100644 --- a/libavfilter/x86/vf_gblur_init.c +++ b/libavfilter/x86/vf_gblur_init.c @@ -24,8 +24,9 @@ #include "libavutil/x86/cpu.h" #include "libavfilter/gblur.h" -void ff_horiz_slice_sse4(float *ptr, int width, int height, int steps, float nu, float bscale); -void ff_horiz_slice_avx2(float *ptr, int width, int height, int steps, float nu, float bscale); +void ff_horiz_slice_sse4(float *ptr, int width, int height, int steps, float nu, float bscale, float *localbuf); +void ff_horiz_slice_avx2(float *ptr, int width, int height, int steps, float nu, float bscale, float *localbuf); +void ff_horiz_slice_avx512(float *ptr, int width, int height, int steps, float nu, float bscale, float *localbuf); void ff_postscale_slice_sse(float *ptr, int length, float postscale, float min, float max); void ff_postscale_slice_avx2(float *ptr, int length, float postscale, float min, float max); @@ -51,12 +52,22 @@ av_cold void ff_gblur_init_x86(GBlurContext *s) s->horiz_slice = ff_horiz_slice_sse4; } if (EXTERNAL_AVX2(cpu_flags)) { - s->horiz_slice = ff_horiz_slice_avx2; s->verti_slice = ff_verti_slice_avx2; } if (EXTERNAL_AVX512(cpu_flags)) { s->postscale_slice = ff_postscale_slice_avx512; s->verti_slice = ff_verti_slice_avx512; } + if (EXTERNAL_AVX2(cpu_flags)) { + s->stride = EXTERNAL_AVX512(cpu_flags) ? 16 : 8; + s->localbuf = av_malloc(s->stride * sizeof(float) * s->planewidth[0] * s->planeheight[0]); + if (!s->localbuf) + return; + + s->horiz_slice = ff_horiz_slice_avx2; + if (EXTERNAL_AVX512(cpu_flags)) { + s->horiz_slice = ff_horiz_slice_avx512; + } + } #endif }