From patchwork Tue Aug 14 07:33:29 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Gagandeep Singh X-Patchwork-Id: 9987 Delivered-To: ffmpegpatchwork@gmail.com Received: by 2002:a02:104:0:0:0:0:0 with SMTP id c4-v6csp4032808jad; Tue, 14 Aug 2018 00:33:50 -0700 (PDT) X-Google-Smtp-Source: AA+uWPzmYI0coTUbOZAGbPJ4o9z1AQ6VQYwwM96HdJ13fEEyuRdMERnUGm262vPVkquH7j86LhCV X-Received: by 2002:a1c:7d47:: with SMTP id y68-v6mr9502234wmc.62.1534232030701; Tue, 14 Aug 2018 00:33:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1534232030; cv=none; d=google.com; s=arc-20160816; b=lhXdIocpy4NKFnhd6pOF6wReijfGgjeEKefMmGRnmMSv71EE4RKUUjl1rz9ZMD2JMR DJr7WgSxXjk9IlrKTq4wt1sxKjVwI+oMN3lKFR3+cs41qUPaiqhYcweVsr6ZY9/eguIM 8VRC9DqiIWMcj2dVwayTLVR45Uy7rBMFMwHzUAptcpUYKk0xf02qoWfFOq62YoPSvlDH u/DSNyOfM+Wp74sGDis6yBXG0BcbtLIkM5PUKg88CIG467i0Fn6CeinVuRcvfrmu9Twp cRA969Uh4OsQBwcYl9J4WS1bAYqDBlhjQZ7RcExVOfpXj85UPY3oTRLBxZ6ayBiHHOQa QZMQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:content-transfer-encoding:mime-version:cc:reply-to :list-subscribe:list-help:list-post:list-archive:list-unsubscribe :list-id:precedence:subject:message-id:date:to:from:dkim-signature :delivered-to:arc-authentication-results; bh=Wh2iUuCE1agoperS1b+Fd8pC0h0Sx5+KeBcs231bdeI=; b=z74nZ91/yvZL27uyHl+VxihVC3NXp5T0O/i8xtmJITVuuSYGOKqeHYga8SqfxBP/ia m9xyBcbwlfsc5P8HsaC1KNOXZhKFiooN+8QG6FZcHMQN0XgXsU0gLzVBdp0tpUPhDBYq 8OJhCxldUzLw6WM2yJ5xtqjY7Sva0OfbwyNNeDufZBUQymrwzO6nc8DcbUOrP6IMoe3p sp6haVDvR+DfPNBRShN11+9dDAuj9gOKbQ5a9bR2x/E/erQqXBQYJh1oaJ76Y/Ge5vuR RzlMIEXWdRQMI7FnAi/IlXpzjBP4xpVV8dS7CqnbIZjwSZTbhGDK3ULl7QwrA6DILjKT S9zw== ARC-Authentication-Results: i=1; mx.google.com; dkim=neutral (body hash did not verify) header.i=@gmail.com header.s=20161025 header.b="SKg/Knii"; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org; dmarc=fail (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from ffbox0-bg.mplayerhq.hu (ffbox0-bg.ffmpeg.org. [79.124.17.100]) by mx.google.com with ESMTP id d6-v6si14274185wrv.432.2018.08.14.00.33.50; Tue, 14 Aug 2018 00:33:50 -0700 (PDT) Received-SPF: pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) client-ip=79.124.17.100; Authentication-Results: mx.google.com; dkim=neutral (body hash did not verify) header.i=@gmail.com header.s=20161025 header.b="SKg/Knii"; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org; dmarc=fail (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: from [127.0.1.1] (localhost [127.0.0.1]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id EFF3A68A552; Tue, 14 Aug 2018 10:33:24 +0300 (EEST) X-Original-To: ffmpeg-devel@ffmpeg.org Delivered-To: ffmpeg-devel@ffmpeg.org Received: from mail-pf1-f175.google.com (mail-pf1-f175.google.com [209.85.210.175]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTPS id 353E5689BAE for ; Tue, 14 Aug 2018 10:33:18 +0300 (EEST) Received: by mail-pf1-f175.google.com with SMTP id u24-v6so8876188pfn.13 for ; Tue, 14 Aug 2018 00:33:42 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id; bh=A8ntRC+wO6MpCaN4ANTGptgVsuWYmKqAPITZYHLNvmA=; b=SKg/KniiKtlwFmrKd4q0tdVzZPgSHpdWJtC4EzEa1GXhsRSTeRTmcvZGhS/J6Nve+S LAHurzsZmz+C9moAHuPy1bjvshTVjRHQN9Qgk04EvrG9gbFmL4bb5bc7TM4lORiI4B3x eilXtVcbyUq5LCqLx/TmNBaq38ist1bKlIFNx1U7asSHPVYdyEQSfbVuNPwnyR6pL6AN uoepNvkJiAmVC0moo9CEqBjeAKoSAY2LBCnLWPRgivaHO0krJChfohPm9aqFng5KPjSw Hw80sJbMfzoXrNXEWRrbFfnURneRrSA0n2NmlI9V7pTQvNYlj1l2wJ+eUGi938yq+M0a pEmg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id; bh=A8ntRC+wO6MpCaN4ANTGptgVsuWYmKqAPITZYHLNvmA=; b=jBTFn7v3PSyUNBoDUhOO75yLQno51F+rP2dI+7OrF9BCfl7F2PK3rMILSICbslMcEz CzyMGdQrCRAmYBhKRaW3wgVmW+zUBph7MbQ89Kq1kGCEr9kI6QfNU6R20bcmh236ZN2P 969t7WcULe1//Wg6onn1ENaRE0OjFw3B3kKEKIL9OSrcDRbirX2S75ny5B3CkZGuvQEI HhDAgJnJqlIdnJV5O2B6bq0Q5SmtIrPaFE5U3hFcgskLEt1LaZG5w40OD0MsxCPyEbFH eNyQ5Ls3s2bZBmaoGwfBrx6FGXriWtDwulCf5szwrJD99wK1U7oUhvs9LFTb6ccuj3jD fEqQ== X-Gm-Message-State: AOUpUlFwNYwY81XtAUukJwqeoBjWBIt+AtNZxgPRfquk84KDJdGRfgEg 5yGfii+/4di0JlVzQVhRCW/q9to6 X-Received: by 2002:a62:ea05:: with SMTP id t5-v6mr22109242pfh.228.1534232019749; Tue, 14 Aug 2018 00:33:39 -0700 (PDT) Received: from gagandeep-Aspire-VX5-591G.cse.iitk.ac.in ([125.17.242.34]) by smtp.googlemail.com with ESMTPSA id d75-v6sm30655632pfd.49.2018.08.14.00.33.37 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 14 Aug 2018 00:33:39 -0700 (PDT) From: Gagandeep Singh To: ffmpeg-devel@ffmpeg.org Date: Tue, 14 Aug 2018 13:03:29 +0530 Message-Id: <20180814073329.7257-1-deepgagan231197@gmail.com> X-Mailer: git-send-email 2.17.1 Subject: [FFmpeg-devel] [GSOC][PATCH 1/3] lavc/cfhd:3d transform decoding for both progressive and interlaced X-BeenThere: ffmpeg-devel@ffmpeg.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: FFmpeg development discussions and patches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Reply-To: FFmpeg development discussions and patches Cc: Gagandeep Singh MIME-Version: 1.0 Errors-To: ffmpeg-devel-bounces@ffmpeg.org Sender: "ffmpeg-devel" --- libavcodec/cfhd.c | 511 ++++++++++++++++++++++++++++++++++++++++------ libavcodec/cfhd.h | 13 +- 2 files changed, 454 insertions(+), 70 deletions(-) diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c index 76838f0869..3906d7a86d 100644 --- a/libavcodec/cfhd.c +++ b/libavcodec/cfhd.c @@ -41,12 +41,15 @@ #define ALPHA_COMPAND_GAIN 9400 enum CFHDParam { + TransformType = 10, ChannelCount = 12, SubbandCount = 14, + Pframe = 19, ImageWidth = 20, ImageHeight = 21, LowpassPrecision = 35, SubbandNumber = 48, + EncodingMethod = 52, Quantization = 53, ChannelNumber = 62, SampleFlags = 68, @@ -64,6 +67,7 @@ static av_cold int cfhd_init(AVCodecContext *avctx) avctx->bits_per_raw_sample = 10; s->avctx = avctx; + s->progressive = 0; return ff_cfhd_init_vlcs(s); } @@ -84,6 +88,10 @@ static void init_peak_table_defaults(CFHDContext *s) static void init_frame_defaults(CFHDContext *s) { + s->sample_type = 0; + s->transform_type = 0; + s->pframe = 0; + s->first_wavelet = 0; s->coded_width = 0; s->coded_height = 0; s->cropped_height = 0; @@ -97,14 +105,15 @@ static void init_frame_defaults(CFHDContext *s) s->pshift = 1; s->codebook = 0; s->difference_coding = 0; - s->progressive = 0; init_plane_defaults(s); init_peak_table_defaults(s); } /* TODO: merge with VLC tables or use LUT */ -static inline int dequant_and_decompand(int level, int quantisation, int codebook) +static inline int dequant_and_decompand(int level, int quantisation, int codebook, int lossless) { + if (lossless) + return level; if (codebook == 0 || codebook == 1) { int64_t abslevel = abs(level); if (level < 264) @@ -193,16 +202,21 @@ static inline void filter(int16_t *output, ptrdiff_t out_stride, } } -static inline void interlaced_vertical_filter(int16_t *output, int16_t *low, int16_t *high, - int width, int linesize, int plane) +static inline void inverse_temporal_filter(int16_t *output, int16_t *low, int16_t *high, + int width, int linesize, int temporal_for_highpass) { int i; int16_t even, odd; for (i = 0; i < width; i++) { even = (low[i] - high[i])/2; odd = (low[i] + high[i])/2; - output[i] = av_clip_uintp2(even, 10); - output[i + linesize] = av_clip_uintp2(odd, 10); + if (!temporal_for_highpass) { + output[i] = av_clip_uintp2(even, 10); + output[i + linesize] = av_clip_uintp2(odd, 10); + } else { + low[i] = even; + high[i] = odd; + } } } static void horiz_filter(int16_t *output, int16_t *low, int16_t *high, @@ -231,9 +245,12 @@ static void free_buffers(CFHDContext *s) for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) { av_freep(&s->plane[i].idwt_buf); av_freep(&s->plane[i].idwt_tmp); - - for (j = 0; j < 9; j++) - s->plane[i].subband[j] = NULL; + if (s->transform_type == 0) + for (j = 0; j < 9; j++) + s->plane[i].subband[j] = NULL; + else + for (j = 0; j < 17; j++) + s->plane[i].subband[j] = NULL; for (j = 0; j < 8; j++) s->plane[i].l_h[j] = NULL; @@ -247,7 +264,7 @@ static int alloc_buffers(AVCodecContext *avctx) CFHDContext *s = avctx->priv_data; int i, j, ret, planes; int chroma_x_shift, chroma_y_shift; - unsigned k; + unsigned k, t; if ((ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height)) < 0) return ret; @@ -261,6 +278,7 @@ static int alloc_buffers(AVCodecContext *avctx) for (i = 0; i < planes; i++) { int w8, h8, w4, h4, w2, h2; + int16_t *frame2; int width = i ? avctx->width >> chroma_x_shift : avctx->width; int height = i ? avctx->height >> chroma_y_shift : avctx->height; ptrdiff_t stride = FFALIGN(width / 8, 8) * 8; @@ -277,28 +295,68 @@ static int alloc_buffers(AVCodecContext *avctx) w2 = w4 * 2; h2 = h4 * 2; - s->plane[i].idwt_buf = - av_mallocz_array(height * stride, sizeof(*s->plane[i].idwt_buf)); - s->plane[i].idwt_tmp = - av_malloc_array(height * stride, sizeof(*s->plane[i].idwt_tmp)); - if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp) - return AVERROR(ENOMEM); - - s->plane[i].subband[0] = s->plane[i].idwt_buf; - s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8; - s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8; - s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8; - s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4; - s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4; - s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4; - s->plane[i].subband[7] = s->plane[i].idwt_buf + 2 * w2 * h2; - s->plane[i].subband[8] = s->plane[i].idwt_buf + 1 * w2 * h2; - s->plane[i].subband[9] = s->plane[i].idwt_buf + 3 * w2 * h2; - - for (j = 0; j < DWT_LEVELS; j++) { - for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++) { - s->plane[i].band[j][k].a_width = w8 << j; - s->plane[i].band[j][k].a_height = h8 << j; + if (s->transform_type == 0) { + s->plane[i].idwt_buf = + av_mallocz_array(height * stride, sizeof(*s->plane[i].idwt_buf)); + s->plane[i].idwt_tmp = + av_malloc_array(height * stride, sizeof(*s->plane[i].idwt_tmp)); + if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp) + return AVERROR(ENOMEM); + } else if (s->transform_type == 2) { + s->plane[i].idwt_buf = + av_mallocz_array(2 * height * stride, sizeof(*s->plane[i].idwt_buf)); + s->plane[i].idwt_tmp = + av_malloc_array(2 * height * stride, sizeof(*s->plane[i].idwt_tmp)); + if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp) + return AVERROR(ENOMEM); + } + + if (s->transform_type == 0) { + s->plane[i].subband[0] = s->plane[i].idwt_buf; + s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8; + s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8; + s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8; + s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4; + s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4; + s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4; + s->plane[i].subband[7] = s->plane[i].idwt_buf + 2 * w2 * h2; + s->plane[i].subband[8] = s->plane[i].idwt_buf + 1 * w2 * h2; + s->plane[i].subband[9] = s->plane[i].idwt_buf + 3 * w2 * h2; + } else if (s->transform_type == 2) { + s->plane[i].subband[0] = s->plane[i].idwt_buf; + s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8; + s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8; + s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8; + s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4; + s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4; + s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4; + frame2 = + s->plane[i].subband[7] = s->plane[i].idwt_buf + 4 * w2 * h2; + s->plane[i].subband[8] = frame2 + 2 * w4 * h4; + s->plane[i].subband[9] = frame2 + 1 * w4 * h4; + s->plane[i].subband[10] = frame2 + 3 * w4 * h4; + s->plane[i].subband[11] = frame2 + 2 * w2 * h2; + s->plane[i].subband[12] = frame2 + 1 * w2 * h2; + s->plane[i].subband[13] = frame2 + 3 * w2 * h2; + s->plane[i].subband[14] = s->plane[i].idwt_buf + 2 * w2 * h2; + s->plane[i].subband[15] = s->plane[i].idwt_buf + 1 * w2 * h2; + s->plane[i].subband[16] = s->plane[i].idwt_buf + 3 * w2 * h2; + } + + if (s->transform_type == 0) { + for (j = 0; j < DWT_LEVELS - 3; j++) { + for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++) { + s->plane[i].band[j][k].a_width = w8 << j; + s->plane[i].band[j][k].a_height = h8 << j; + } + } + } else if (s->transform_type == 2) { + for (j = 0; j < DWT_LEVELS; j++) { + t = j < 1 ? 0 : (j < 3 ? 1 : 2); + for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[0]); k++) { + s->plane[i].band[j][k].a_width = w8 << t; + s->plane[i].band[j][k].a_height = h8 << t; + } } } @@ -311,6 +369,11 @@ static int alloc_buffers(AVCodecContext *avctx) // s->plane[i].l_h[5] = ll1; s->plane[i].l_h[6] = s->plane[i].idwt_tmp; s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2; + if (s->transform_type == 2) { + frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2; + s->plane[i].l_h[8] = frame2; + s->plane[i].l_h[9] = frame2 + 2 * w2 * h2; + } } s->a_height = s->coded_height; @@ -349,6 +412,9 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } else if (tag == SampleFlags) { av_log(avctx, AV_LOG_DEBUG, "Progressive?%"PRIu16"\n", data); s->progressive = data & 0x0001; + } else if (tag == Pframe) { + s->pframe = 1; + av_log(avctx, AV_LOG_DEBUG, "Frame type %"PRIu16"\n", data); } else if (tag == ImageWidth) { av_log(avctx, AV_LOG_DEBUG, "Width %"PRIu16"\n", data); s->coded_width = data; @@ -373,7 +439,7 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } } else if (tag == SubbandCount) { av_log(avctx, AV_LOG_DEBUG, "Subband Count: %"PRIu16"\n", data); - if (data != SUBBAND_COUNT) { + if (data != 10 && data != 17) { av_log(avctx, AV_LOG_ERROR, "Subband Count of %"PRIu16" is unsupported\n", data); ret = AVERROR_PATCHWELCOME; break; @@ -405,7 +471,7 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } else if (tag == 51) { av_log(avctx, AV_LOG_DEBUG, "Subband number actual %"PRIu16"\n", data); s->subband_num_actual = data; - if (s->subband_num_actual >= 10) { + if (s->subband_num_actual >= 17 && s->subband_num_actual != 255) { av_log(avctx, AV_LOG_ERROR, "Invalid subband number actual\n"); ret = AVERROR(EINVAL); break; @@ -420,9 +486,15 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, s->prescale_shift[1] = (data >> 3) & 0x7; s->prescale_shift[2] = (data >> 6) & 0x7; av_log(avctx, AV_LOG_DEBUG, "Prescale shift (VC-5): %x\n", data); + } else if (tag == EncodingMethod) { + s->encode_method = data; + av_log(avctx, AV_LOG_DEBUG, "Encode Method for Subband %d : %x\n",s->subband_num_actual, data); } else if (tag == 27) { av_log(avctx, AV_LOG_DEBUG, "Lowpass width %"PRIu16"\n", data); - if (data < 3 || data > s->plane[s->channel_num].band[0][0].a_width) { + if (s->coded_width == 0){ + s->coded_width = data << 3; + } + if (data < 3) { av_log(avctx, AV_LOG_ERROR, "Invalid lowpass width\n"); ret = AVERROR(EINVAL); break; @@ -431,20 +503,21 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, s->plane[s->channel_num].band[0][0].stride = data; } else if (tag == 28) { av_log(avctx, AV_LOG_DEBUG, "Lowpass height %"PRIu16"\n", data); - if (data < 3 || data > s->plane[s->channel_num].band[0][0].a_height) { + if (s->coded_height == 0) + s->coded_height = data << 3; + if (data < 3) { av_log(avctx, AV_LOG_ERROR, "Invalid lowpass height\n"); ret = AVERROR(EINVAL); break; } s->plane[s->channel_num].band[0][0].height = data; - } else if (tag == 1) + } else if (tag == 1) { + s->sample_type = data; + if (data == 2) + s->pframe = 1; av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data); - else if (tag == 10) { - if (data != 0) { - avpriv_report_missing_feature(avctx, "Transform type of %"PRIu16, data); - ret = AVERROR_PATCHWELCOME; - break; - } + } else if (tag == 10) { + s->transform_type = data; av_log(avctx, AV_LOG_DEBUG, "Transform-type? %"PRIu16"\n", data); } else if (abstag >= 0x4000 && abstag <= 0x40ff) { if (abstag == 0x4001) @@ -551,8 +624,8 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, av_log(avctx, AV_LOG_DEBUG, "Unknown tag %i data %x\n", tag, data); /* Some kind of end of header tag */ - if (tag == 4 && data == 0x1a4a && s->coded_width && s->coded_height && - s->coded_format != AV_PIX_FMT_NONE) { + if (((tag == 4 && (data == 0x1a4a || data == 0xf0f)) || s->sample_type == 1) && s->coded_width && s->coded_height && + s->coded_format != AV_PIX_FMT_NONE && s->sample_type != 3 && s->sample_type != 6) { if (s->a_width != s->coded_width || s->a_height != s->coded_height || s->a_format != s->coded_format) { free_buffers(s); @@ -619,8 +692,7 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, av_log(avctx, AV_LOG_DEBUG, "Lowpass coefficients %d\n", lowpass_width * lowpass_height); } - - if (tag == 55 && s->subband_num_actual != 255 && s->a_width && s->a_height) { + if ((tag == 55 || tag == 82) && s->a_width && s->a_height) { int highpass_height = s->plane[s->channel_num].band[s->level][s->subband_num].height; int highpass_width = s->plane[s->channel_num].band[s->level][s->subband_num].width; int highpass_a_width = s->plane[s->channel_num].band[s->level][s->subband_num].a_width; @@ -638,18 +710,24 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } if (highpass_height > highpass_a_height || highpass_width > highpass_a_width || a_expected < highpass_height * (uint64_t)highpass_stride) { - av_log(avctx, AV_LOG_ERROR, "Too many highpass coefficients\n"); - ret = AVERROR(EINVAL); - goto end; + if (s->subband_num_actual != 255) { + av_log(avctx, AV_LOG_ERROR, "Too many highpass coefficients\n"); + ret = AVERROR(EINVAL); + goto end; + } } expected = highpass_height * highpass_stride; av_log(avctx, AV_LOG_DEBUG, "Start subband coeffs plane %i level %i codebook %i expected %i\n", s->channel_num, s->level, s->codebook, expected); init_get_bits(&s->gb, gb.buffer, bytestream2_get_bytes_left(&gb) * 8); + if (s->subband_num_actual == 255) { + expected = 0; + goto finish; + } { OPEN_READER(re, &s->gb); - if (!s->codebook) { + if (!s->codebook && !(s->transform_type == 2 && s->subband_num_actual == 7)) { while (1) { UPDATE_CACHE(re, &s->gb); GET_RL_VLC(level, run, re, &s->gb, s->table_9_rl_vlc, @@ -664,9 +742,14 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, if (count > expected) break; - coeff = dequant_and_decompand(level, s->quantisation, 0); + coeff = dequant_and_decompand(level, s->quantisation, 0, (s->sample_type == 2 || s->sample_type == 3) && s->pframe && s->subband_num_actual == 7 && s->encode_method == 5); for (i = 0; i < run; i++) - *coeff_data++ = coeff; + if (tag != 82) + *coeff_data++ = coeff; + else { + *coeff_data |= coeff << 8; + *coeff_data++ *= s->quantisation; + } } if (s->peak.level) peak_table(coeff_data - expected, &s->peak, expected); @@ -688,9 +771,14 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, if (count > expected) break; - coeff = dequant_and_decompand(level, s->quantisation, s->codebook); + coeff = dequant_and_decompand(level, s->quantisation, s->codebook, (s->sample_type == 2 || s->sample_type == 3) && s->pframe && s->subband_num_actual == 7 && s->encode_method == 5); for (i = 0; i < run; i++) - *coeff_data++ = coeff; + if (tag != 82) + *coeff_data++ = coeff; + else { + *coeff_data |= coeff << 8; + *coeff_data++ *= s->quantisation; + } } if (s->peak.level) peak_table(coeff_data - expected, &s->peak, expected); @@ -701,12 +789,12 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, CLOSE_READER(re, &s->gb); } - if (count > expected) { + if (count > expected && s->subband_num_actual != 255) { av_log(avctx, AV_LOG_ERROR, "Escape codeword not found, probably corrupt data\n"); ret = AVERROR(EINVAL); goto end; } - + finish: bytes = FFALIGN(AV_CEIL_RSHIFT(get_bits_count(&s->gb), 3), 4); if (bytes > bytestream2_get_bytes_left(&gb)) { av_log(avctx, AV_LOG_ERROR, "Bitstream overread error\n"); @@ -726,14 +814,15 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } } } - - if (!s->a_width || !s->a_height || s->a_format == AV_PIX_FMT_NONE || - s->coded_width || s->coded_height || s->coded_format != AV_PIX_FMT_NONE) { + //disabled to run mountain sample file +#if 0 + if ((!s->a_width || !s->a_height || s->a_format == AV_PIX_FMT_NONE || + s->coded_width || s->coded_height || s->coded_format != AV_PIX_FMT_NONE) && s->sample_type != 1) { av_log(avctx, AV_LOG_ERROR, "Invalid dimensions\n"); ret = AVERROR(EINVAL); goto end; } - +#endif if (!got_buffer) { av_log(avctx, AV_LOG_ERROR, "No end of header tag found\n"); ret = AVERROR(EINVAL); @@ -741,8 +830,9 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } planes = av_pix_fmt_count_planes(avctx->pix_fmt); + if (s->transform_type == 0 && s->sample_type != 1) { for (plane = 0; plane < planes && !ret; plane++) { - /* level 1 */ + /* level 1 */ int lowpass_height = s->plane[plane].band[0][0].height; int lowpass_width = s->plane[plane].band[0][0].width; int highpass_stride = s->plane[plane].band[0][1].stride; @@ -921,14 +1011,305 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, low = s->plane[plane].l_h[6]; high = s->plane[plane].l_h[7]; for (i = 0; i < lowpass_height; i++) { - interlaced_vertical_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, act_plane); + inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); low += lowpass_width * 2; high += lowpass_width * 2; dst += pic->linesize[act_plane]; } } } + //this is the serial version on ip sample decoding so buffers allocated using alloc_buffers() are not freed, + //so the stored decoded coefficients data is used for generating the second frame once empty packet is passed in sample_type = 1 + } else if (s->transform_type == 2 && s->sample_type != 1) { + for (plane = 0; plane < planes && !ret; plane++) { + int lowpass_height = s->plane[plane].band[0][0].height; + int lowpass_width = s->plane[plane].band[0][0].width; + int highpass_stride = s->plane[plane].band[0][1].stride; + int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane; + int16_t *low, *high, *output, *dst; + + if (lowpass_height > s->plane[plane].band[0][0].a_height || lowpass_width > s->plane[plane].band[0][0].a_width || + !highpass_stride || s->plane[plane].band[0][1].width > s->plane[plane].band[0][1].a_width) { + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n"); + ret = AVERROR(EINVAL); + goto end; + } + + av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); + + low = s->plane[plane].subband[0]; + high = s->plane[plane].subband[2]; + output = s->plane[plane].l_h[0]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].subband[1]; + high = s->plane[plane].subband[3]; + output = s->plane[plane].l_h[1]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].l_h[0]; + high = s->plane[plane].l_h[1]; + output = s->plane[plane].subband[0]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + if (s->bpc == 12) { + output = s->plane[plane].subband[0]; + for (i = 0; i < lowpass_height * 2; i++) { + for (j = 0; j < lowpass_width * 2; j++) + output[j] *= 4; + + output += lowpass_width * 2; + } + } + + lowpass_height = s->plane[plane].band[1][1].height; + lowpass_width = s->plane[plane].band[1][1].width; + highpass_stride = s->plane[plane].band[1][1].stride; + + if (lowpass_height > s->plane[plane].band[1][1].a_height || lowpass_width > s->plane[plane].band[1][1].a_width || + !highpass_stride || s->plane[plane].band[1][1].width > s->plane[plane].band[1][1].a_width) { + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n"); + ret = AVERROR(EINVAL); + goto end; + } + + av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); + + low = s->plane[plane].subband[0]; + high = s->plane[plane].subband[5]; + output = s->plane[plane].l_h[3]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + low = s->plane[plane].subband[4]; + high = s->plane[plane].subband[6]; + output = s->plane[plane].l_h[4]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].l_h[3]; + high = s->plane[plane].l_h[4]; + output = s->plane[plane].subband[0]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + output = s->plane[plane].subband[0]; + for (i = 0; i < lowpass_height * 2; i++) { + for (j = 0; j < lowpass_width * 2; j++) + output[j] *= 4; + output += lowpass_width * 2; + } + + low = s->plane[plane].subband[7]; + high = s->plane[plane].subband[9]; + output = s->plane[plane].l_h[3]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].subband[8]; + high = s->plane[plane].subband[10]; + output = s->plane[plane].l_h[4]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].l_h[3]; + high = s->plane[plane].l_h[4]; + output = s->plane[plane].subband[7]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + lowpass_height = s->plane[plane].band[4][1].height; + lowpass_width = s->plane[plane].band[4][1].width; + highpass_stride = s->plane[plane].band[4][1].stride; + av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); + + if (lowpass_height > s->plane[plane].band[4][1].a_height || lowpass_width > s->plane[plane].band[4][1].a_width || + !highpass_stride || s->plane[plane].band[4][1].width > s->plane[plane].band[4][1].a_width) { + av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n"); + ret = AVERROR(EINVAL); + goto end; + } + + low = s->plane[plane].subband[0]; + high = s->plane[plane].subband[7]; + output = s->plane[plane].subband[0]; + for (i = 0; i < lowpass_height; i++) { + inverse_temporal_filter(output, low, high, lowpass_width, 4 * lowpass_width * lowpass_height, 1); + low += lowpass_width; + high += lowpass_width; + } + if (s->progressive) { + low = s->plane[plane].subband[0]; + high = s->plane[plane].subband[15]; + output = s->plane[plane].l_h[6]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].subband[14]; + high = s->plane[plane].subband[16]; + output = s->plane[plane].l_h[7]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].subband[7]; + high = s->plane[plane].subband[12]; + output = s->plane[plane].l_h[8]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->plane[plane].subband[11]; + high = s->plane[plane].subband[13]; + output = s->plane[plane].l_h[9]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + dst = (int16_t *)pic->data[act_plane]; + low = s->plane[plane].l_h[6]; + high = s->plane[plane].l_h[7]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(dst, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + dst += pic->linesize[act_plane] / 2; + } + } else { + pic->interlaced_frame = 1; + low = s->plane[plane].subband[0]; + high = s->plane[plane].subband[14]; + output = s->plane[plane].l_h[6]; + for (i = 0; i < lowpass_height; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + low = s->plane[plane].subband[15]; + high = s->plane[plane].subband[16]; + output = s->plane[plane].l_h[7]; + for (i = 0; i < lowpass_height; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + low = s->plane[plane].subband[7]; + high = s->plane[plane].subband[11]; + output = s->plane[plane].l_h[8]; + for (i = 0; i < lowpass_height; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + low = s->plane[plane].subband[12]; + high = s->plane[plane].subband[13]; + output = s->plane[plane].l_h[9]; + for (i = 0; i < lowpass_height; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + + dst = (int16_t *)pic->data[act_plane]; + low = s->plane[plane].l_h[6]; + high = s->plane[plane].l_h[7]; + for (i = 0; i < lowpass_height; i++) { + inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); + low += lowpass_width * 2; + high += lowpass_width * 2; + dst += pic->linesize[act_plane]; + } + } + } + } else if (s->sample_type == 1) { + int16_t *low, *high, *dst; + int lowpass_height, lowpass_width; + for (plane = 0; plane < planes && !ret; plane++) { + int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane; + lowpass_height = s->plane[plane].band[4][1].height; + lowpass_width = s->plane[plane].band[4][1].width; + if (s->progressive) { + dst = (int16_t *)pic->data[act_plane]; + low = s->plane[plane].l_h[8]; + high = s->plane[plane].l_h[9]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(dst, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + dst += pic->linesize[act_plane] / 2; + } + } else { + dst = (int16_t *)pic->data[act_plane]; + low = s->plane[plane].l_h[8]; + high = s->plane[plane].l_h[9]; + for (i = 0; i < lowpass_height; i++) { + inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); + low += lowpass_width * 2; + high += lowpass_width * 2; + dst += pic->linesize[act_plane]; + } + } + } + } end: if (ret < 0) @@ -961,6 +1342,6 @@ AVCodec ff_cfhd_decoder = { .init = cfhd_init, .close = cfhd_close, .decode = cfhd_decode, - .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS, - .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP, + .capabilities = AV_CODEC_CAP_DR1, + .caps_internal = FF_CODEC_CAP_INIT_CLEANUP, }; diff --git a/libavcodec/cfhd.h b/libavcodec/cfhd.h index 4f2c82d8bc..047c0f2028 100644 --- a/libavcodec/cfhd.h +++ b/libavcodec/cfhd.h @@ -31,15 +31,14 @@ #include "vlc.h" #define VLC_BITS 9 -#define SUBBAND_COUNT 10 - +#define SUBBAND_COUNT 17 typedef struct CFHD_RL_VLC_ELEM { int16_t level; int8_t len; uint16_t run; } CFHD_RL_VLC_ELEM; -#define DWT_LEVELS 3 +#define DWT_LEVELS 6 typedef struct SubBand { int level; @@ -64,7 +63,7 @@ typedef struct Plane { /* TODO: merge this into SubBand structure */ int16_t *subband[SUBBAND_COUNT]; - int16_t *l_h[8]; + int16_t *l_h[10]; SubBand band[DWT_LEVELS][4]; } Plane; @@ -83,9 +82,13 @@ typedef struct CFHDContext { CFHD_RL_VLC_ELEM table_18_rl_vlc[4572]; VLC vlc_18; - GetBitContext gb; + int sample_type; + int transform_type; + int encode_method; + int first_wavelet; + int pframe; int coded_width; int coded_height; int cropped_height; -- 2.17.1 From d6ac8a4ee2f08a6d3cab7e7500d740c661062c9f Mon Sep 17 00:00:00 2001 From: Gagandeep Singh Date: Tue, 14 Aug 2018 00:20:06 +0530 Subject: [GSOC][FFmpeg-devel][PATCH 2/3] lavc/cfhd:corrected decompanding for table 9 used in cfhd --- libavcodec/cfhd.c | 25 ++++++++++++++++++++++++- 1 file changed, 24 insertions(+), 1 deletion(-) diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c index 3906d7a86d..2c538f0bbd 100644 --- a/libavcodec/cfhd.c +++ b/libavcodec/cfhd.c @@ -114,7 +114,30 @@ static inline int dequant_and_decompand(int level, int quantisation, int codeboo { if (lossless) return level; - if (codebook == 0 || codebook == 1) { + if (codebook == 0) { + if (level >= 40 && level < 264) { + if (level >= 54) { + level -= 54; + level <<= 2; + level += 54; + } + level -= 40; + level <<= 2; + level += 40; + } else if (level <= -40) { + level = -level; + if (level >= 54) { + level -= 54; + level <<= 2; + level += 54; + } + level -= 40; + level <<= 2; + level += 40; + level = -level; + } + return level * quantisation; + } else if (codebook == 1) { int64_t abslevel = abs(level); if (level < 264) return (abslevel + ((768 * abslevel * abslevel * abslevel) / (255 * 255 * 255))) * -- 2.17.1 From fa23549c61a6d8413cdc79c570376c53795a6ff1 Mon Sep 17 00:00:00 2001 From: Gagandeep Singh Date: Tue, 14 Aug 2018 12:43:20 +0530 Subject: [GSOC][FFmpeg-devel][PATCH 3/3] lavc/cfhd:frame threading support for 3d transform progressive and interlaced samples --- libavcodec/cfhd.c | 378 ++++++++++++++++++++++++++++------------------ libavcodec/cfhd.h | 8 +- 2 files changed, 242 insertions(+), 144 deletions(-) diff --git a/libavcodec/cfhd.c b/libavcodec/cfhd.c index 2c538f0bbd..7c298056ca 100644 --- a/libavcodec/cfhd.c +++ b/libavcodec/cfhd.c @@ -63,13 +63,23 @@ enum CFHDParam { static av_cold int cfhd_init(AVCodecContext *avctx) { + int ret; + CFHDContext *s = avctx->priv_data; + if (!avctx->internal->is_copy) { + avctx->internal->allocate_progress = 1; + ret = ff_cfhd_init_vlcs(s); + } else + ret = 0; avctx->bits_per_raw_sample = 10; s->avctx = avctx; s->progressive = 0; + s->i_frame.f = av_frame_alloc(); + s->p_frame.f = av_frame_alloc(); - return ff_cfhd_init_vlcs(s); + + return ret; } static void init_plane_defaults(CFHDContext *s) @@ -268,15 +278,18 @@ static void free_buffers(CFHDContext *s) for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) { av_freep(&s->plane[i].idwt_buf); av_freep(&s->plane[i].idwt_tmp); - if (s->transform_type == 0) + if (s->transform_type == 0) { for (j = 0; j < 9; j++) s->plane[i].subband[j] = NULL; - else + for (j = 0; j < 8; j++) + s->plane[i].l_h[j] = NULL; + } + else { for (j = 0; j < 17; j++) s->plane[i].subband[j] = NULL; - - for (j = 0; j < 8; j++) - s->plane[i].l_h[j] = NULL; + for (j = 0; j < 12; j++) + s->plane[i].l_h[j] = NULL; + } } s->a_height = 0; s->a_width = 0; @@ -394,8 +407,10 @@ static int alloc_buffers(AVCodecContext *avctx) s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2; if (s->transform_type == 2) { frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2; - s->plane[i].l_h[8] = frame2; - s->plane[i].l_h[9] = frame2 + 2 * w2 * h2; + s->plane[i].l_h[8] = frame2; + s->plane[i].l_h[9] = frame2 + 2 * w4 * h4; + s->plane[i].l_h[10] = frame2; + s->plane[i].l_h[11] = frame2 + 2 * w2 * h2; } } @@ -406,14 +421,28 @@ static int alloc_buffers(AVCodecContext *avctx) return 0; } +static int update_thread_context(AVCodecContext *dst, const AVCodecContext *src) +{ + CFHDContext *csrc = src->priv_data; + CFHDContext *cdst = dst->priv_data; + cdst->transform_type = csrc->transform_type; + if (csrc->sample_type != 1 && csrc->transform_type != 0) { + cdst->progressive = csrc->progressive; + cdst->picture = &csrc->p_frame; + cdst->connection = &csrc->i_frame; + cdst->buffers = csrc->plane; + } + + return 0; +} + static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt) { CFHDContext *s = avctx->priv_data; GetByteContext gb; ThreadFrame frame = { .f = data }; - AVFrame *pic = data; - int ret = 0, i, j, planes, plane, got_buffer = 0; + int ret = 0, i, j, planes, plane, got_buffer = 0, progress1 = 1, progress2 = 1; int16_t *coeff_data; s->coded_format = AV_PIX_FMT_YUV422P10; @@ -537,7 +566,9 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, } else if (tag == 1) { s->sample_type = data; if (data == 2) - s->pframe = 1; + s->pframe = 1; + else if (data == 1) + s->transform_type = 2; av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data); } else if (tag == 10) { s->transform_type = data; @@ -657,21 +688,54 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, return ret; } } - ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height); - if (ret < 0) - return ret; - if (s->cropped_height) - avctx->height = s->cropped_height; - frame.f->width = - frame.f->height = 0; - - if ((ret = ff_thread_get_buffer(avctx, &frame, 0)) < 0) - return ret; - + if (s->transform_type == 2) { + if (s->sample_type != 1) { + s->picture = &s->i_frame; + s->connection = &s->p_frame; + s->buffers = s->plane; + } + ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height); + if (ret < 0) + return ret; + if (s->sample_type != 1) { + if (s->i_frame.f->data[0]) + ff_thread_release_buffer(avctx, &s->i_frame); + if (s->p_frame.f->data[0]) + ff_thread_release_buffer(avctx, &s->p_frame); + av_frame_copy_props(s->i_frame.f, frame.f); + av_frame_copy_props(s->p_frame.f, frame.f); + if (s->cropped_height) + avctx->height = s->cropped_height; + s->picture->f->width = + s->picture->f->height = 0; + s->connection->f->width = + s->connection->f->height = 0; + if ((ret = ff_thread_get_buffer(avctx, s->picture, 0)) < 0) + return ret; + if ((ret = ff_thread_get_buffer(avctx, s->connection, 0)) < 0) + return ret; + } + } else { + s->picture = &s->i_frame; + s->buffers = s->plane; + if (s->picture->f->data[0]) + ff_thread_release_buffer(avctx, s->picture); + av_frame_copy_props(s->i_frame.f, frame.f); + ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height); + if (ret < 0) + return ret; + if (s->cropped_height) + avctx->height = s->cropped_height; + s->picture->f->width = + s->picture->f->height = 0; + if ((ret = ff_thread_get_buffer(avctx, s->picture, 0)) < 0) + return ret; + } s->coded_width = 0; s->coded_height = 0; s->coded_format = AV_PIX_FMT_NONE; got_buffer = 1; + ff_thread_finish_setup(avctx); } coeff_data = s->plane[s->channel_num].subband[s->subband_num_actual]; @@ -835,6 +899,8 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, &coeff_data[(highpass_height - 1) * highpass_stride], highpass_stride * sizeof(*coeff_data)); } + if (s->transform_type == 2 && s->subband_num_actual == 10) + ff_thread_report_progress(s->picture, progress1 += 1, 0); } } //disabled to run mountain sample file @@ -975,7 +1041,6 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, ret = AVERROR(EINVAL); goto end; } - av_log(avctx, AV_LOG_DEBUG, "Level 3 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); if (s->progressive) { low = s->plane[plane].subband[0]; @@ -998,18 +1063,18 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, output++; } - dst = (int16_t *)pic->data[act_plane]; + dst = (int16_t *)s->picture->f->data[act_plane]; low = s->plane[plane].l_h[6]; high = s->plane[plane].l_h[7]; for (i = 0; i < lowpass_height * 2; i++) { horiz_filter_clip(dst, low, high, lowpass_width, s->bpc); low += lowpass_width; high += lowpass_width; - dst += pic->linesize[act_plane] / 2; + dst += s->picture->f->linesize[act_plane] / 2; } } else { - av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d", pic->interlaced_frame); - pic->interlaced_frame = 1; + av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d", s->picture->f->interlaced_frame); + s->picture->f->interlaced_frame = 1; low = s->plane[plane].subband[0]; high = s->plane[plane].subband[7]; output = s->plane[plane].l_h[6]; @@ -1030,23 +1095,23 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, output += lowpass_width * 2; } - dst = (int16_t *)pic->data[act_plane]; + dst = (int16_t *)s->picture->f->data[act_plane]; low = s->plane[plane].l_h[6]; high = s->plane[plane].l_h[7]; for (i = 0; i < lowpass_height; i++) { - inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); + inverse_temporal_filter(dst, low, high, lowpass_width * 2, s->picture->f->linesize[act_plane]/2, 0); low += lowpass_width * 2; high += lowpass_width * 2; - dst += pic->linesize[act_plane]; + dst += s->picture->f->linesize[act_plane]; } } } - //this is the serial version on ip sample decoding so buffers allocated using alloc_buffers() are not freed, - //so the stored decoded coefficients data is used for generating the second frame once empty packet is passed in sample_type = 1 + av_frame_ref(frame.f, s->picture->f); + ff_thread_report_progress(s->picture, INT_MAX, 0); } else if (s->transform_type == 2 && s->sample_type != 1) { for (plane = 0; plane < planes && !ret; plane++) { - int lowpass_height = s->plane[plane].band[0][0].height; - int lowpass_width = s->plane[plane].band[0][0].width; + int lowpass_height = s->plane[plane].band[0][1].height; + int lowpass_width = s->plane[plane].band[0][1].width; int highpass_stride = s->plane[plane].band[0][1].stride; int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane; int16_t *low, *high, *output, *dst; @@ -1058,8 +1123,6 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, goto end; } - av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); - low = s->plane[plane].subband[0]; high = s->plane[plane].subband[2]; output = s->plane[plane].l_h[0]; @@ -1110,8 +1173,6 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, goto end; } - av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); - low = s->plane[plane].subband[0]; high = s->plane[plane].subband[5]; output = s->plane[plane].l_h[3]; @@ -1149,40 +1210,9 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, output += lowpass_width * 2; } - low = s->plane[plane].subband[7]; - high = s->plane[plane].subband[9]; - output = s->plane[plane].l_h[3]; - for (i = 0; i < lowpass_width; i++) { - vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); - low++; - high++; - output++; - } - - low = s->plane[plane].subband[8]; - high = s->plane[plane].subband[10]; - output = s->plane[plane].l_h[4]; - for (i = 0; i < lowpass_width; i++) { - vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); - low++; - high++; - output++; - } - - low = s->plane[plane].l_h[3]; - high = s->plane[plane].l_h[4]; - output = s->plane[plane].subband[7]; - for (i = 0; i < lowpass_height * 2; i++) { - horiz_filter(output, low, high, lowpass_width); - low += lowpass_width; - high += lowpass_width; - output += lowpass_width * 2; - } - lowpass_height = s->plane[plane].band[4][1].height; lowpass_width = s->plane[plane].band[4][1].width; highpass_stride = s->plane[plane].band[4][1].stride; - av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride); if (lowpass_height > s->plane[plane].band[4][1].a_height || lowpass_width > s->plane[plane].band[4][1].a_width || !highpass_stride || s->plane[plane].band[4][1].width > s->plane[plane].band[4][1].a_width) { @@ -1190,7 +1220,7 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, ret = AVERROR(EINVAL); goto end; } - + ff_thread_await_progress(s->connection, progress2 += 1, 0); low = s->plane[plane].subband[0]; high = s->plane[plane].subband[7]; output = s->plane[plane].subband[0]; @@ -1199,6 +1229,7 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, low += lowpass_width; high += lowpass_width; } + ff_thread_report_progress(s->picture, progress1 += 1, 0); if (s->progressive) { low = s->plane[plane].subband[0]; high = s->plane[plane].subband[15]; @@ -1220,37 +1251,17 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, output++; } - low = s->plane[plane].subband[7]; - high = s->plane[plane].subband[12]; - output = s->plane[plane].l_h[8]; - for (i = 0; i < lowpass_width; i++) { - vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); - low++; - high++; - output++; - } - - low = s->plane[plane].subband[11]; - high = s->plane[plane].subband[13]; - output = s->plane[plane].l_h[9]; - for (i = 0; i < lowpass_width; i++) { - vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); - low++; - high++; - output++; - } - - dst = (int16_t *)pic->data[act_plane]; + dst = (int16_t *)s->picture->f->data[act_plane]; low = s->plane[plane].l_h[6]; high = s->plane[plane].l_h[7]; for (i = 0; i < lowpass_height * 2; i++) { - horiz_filter(dst, low, high, lowpass_width); + horiz_filter_clip(dst, low, high, lowpass_width, s->bpc); low += lowpass_width; high += lowpass_width; - dst += pic->linesize[act_plane] / 2; + dst += s->picture->f->linesize[act_plane] / 2; } } else { - pic->interlaced_frame = 1; + s->picture->f->interlaced_frame = 1; low = s->plane[plane].subband[0]; high = s->plane[plane].subband[14]; output = s->plane[plane].l_h[6]; @@ -1271,67 +1282,137 @@ static int cfhd_decode(AVCodecContext *avctx, void *data, int *got_frame, output += lowpass_width * 2; } - low = s->plane[plane].subband[7]; - high = s->plane[plane].subband[11]; - output = s->plane[plane].l_h[8]; - for (i = 0; i < lowpass_height; i++) { - horiz_filter(output, low, high, lowpass_width); - low += lowpass_width; - high += lowpass_width; - output += lowpass_width * 2; - } - - low = s->plane[plane].subband[12]; - high = s->plane[plane].subband[13]; - output = s->plane[plane].l_h[9]; - for (i = 0; i < lowpass_height; i++) { - horiz_filter(output, low, high, lowpass_width); - low += lowpass_width; - high += lowpass_width; - output += lowpass_width * 2; - } - - - dst = (int16_t *)pic->data[act_plane]; + dst = (int16_t *)s->picture->f->data[act_plane]; low = s->plane[plane].l_h[6]; high = s->plane[plane].l_h[7]; for (i = 0; i < lowpass_height; i++) { - inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); + inverse_temporal_filter(dst, low, high, lowpass_width * 2, s->picture->f->linesize[act_plane]/2, 0); low += lowpass_width * 2; high += lowpass_width * 2; - dst += pic->linesize[act_plane]; + dst += s->picture->f->linesize[act_plane]; } } } + ff_thread_report_progress(s->picture, INT_MAX, 0); + ff_thread_await_progress(s->connection, INT_MAX, 0); + av_frame_ref(frame.f, s->picture->f); } else if (s->sample_type == 1) { - int16_t *low, *high, *dst; - int lowpass_height, lowpass_width; + int16_t *low, *high, *dst, *output; + int lowpass_height, lowpass_width, highpass_stride, act_plane; + progress1 = 1, progress2 = 1; for (plane = 0; plane < planes && !ret; plane++) { - int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane; - lowpass_height = s->plane[plane].band[4][1].height; - lowpass_width = s->plane[plane].band[4][1].width; + ff_thread_await_progress(s->connection, progress1 += 1, 0); + // highpass inverse for temporal + lowpass_height = s->buffers[plane].band[1][1].a_height; + lowpass_width = s->buffers[plane].band[1][1].a_width; + highpass_stride = s->buffers[plane].band[1][1].a_width; + + low = s->buffers[plane].subband[7]; + high = s->buffers[plane].subband[9]; + output = s->buffers[plane].l_h[8]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->buffers[plane].subband[8]; + high = s->buffers[plane].subband[10]; + output = s->buffers[plane].l_h[9]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->buffers[plane].l_h[8]; + high = s->buffers[plane].l_h[9]; + output = s->buffers[plane].subband[7]; + for (i = 0; i < lowpass_height * 2; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + ff_thread_report_progress(s->picture, progress2 += 1, 0); + } + for (plane = 0; plane < planes && !ret; plane++) { + ff_thread_await_progress(s->connection, progress1 += 1, 0); + + act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane; + lowpass_height = s->buffers[plane].band[4][1].a_height; + lowpass_width = s->buffers[plane].band[4][1].a_width; + highpass_stride = s->buffers[plane].band[4][1].a_width; + if (s->progressive) { - dst = (int16_t *)pic->data[act_plane]; - low = s->plane[plane].l_h[8]; - high = s->plane[plane].l_h[9]; + low = s->buffers[plane].subband[7]; + high = s->buffers[plane].subband[12]; + output = s->buffers[plane].l_h[10]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, lowpass_width, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + low = s->buffers[plane].subband[11]; + high = s->buffers[plane].subband[13]; + output = s->buffers[plane].l_h[11]; + for (i = 0; i < lowpass_width; i++) { + vert_filter(output, lowpass_width, low, highpass_stride, high, highpass_stride, lowpass_height); + low++; + high++; + output++; + } + + dst = (int16_t *)s->picture->f->data[act_plane]; + low = s->buffers[plane].l_h[10]; + high = s->buffers[plane].l_h[11]; for (i = 0; i < lowpass_height * 2; i++) { - horiz_filter(dst, low, high, lowpass_width); + horiz_filter_clip(dst, low, high, lowpass_width, s->bpc); low += lowpass_width; high += lowpass_width; - dst += pic->linesize[act_plane] / 2; + dst += s->picture->f->linesize[act_plane] / 2; } } else { - dst = (int16_t *)pic->data[act_plane]; - low = s->plane[plane].l_h[8]; - high = s->plane[plane].l_h[9]; + av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d", s->picture->f->interlaced_frame); + s->picture->f->interlaced_frame = 1; + low = s->buffers[plane].subband[7]; + high = s->buffers[plane].subband[11]; + output = s->buffers[plane].l_h[10]; + for (i = 0; i < lowpass_height; i++) { + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + low = s->buffers[plane].subband[12]; + high = s->buffers[plane].subband[13]; + output = s->buffers[plane].l_h[11]; for (i = 0; i < lowpass_height; i++) { - inverse_temporal_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, 0); + horiz_filter(output, low, high, lowpass_width); + low += lowpass_width; + high += lowpass_width; + output += lowpass_width * 2; + } + + dst = (int16_t *)s->picture->f->data[act_plane]; + low = s->buffers[plane].l_h[10]; + high = s->buffers[plane].l_h[11]; + for (i = 0; i < lowpass_height; i++) { + inverse_temporal_filter(dst, low, high, lowpass_width * 2, s->picture->f->linesize[act_plane]/2, 0); low += lowpass_width * 2; high += lowpass_width * 2; - dst += pic->linesize[act_plane]; + dst += s->picture->f->linesize[act_plane]; } } } + ff_thread_report_progress(s->picture, INT_MAX, 0); + ff_thread_await_progress(s->connection, INT_MAX, 0); + av_frame_ref(frame.f, s->picture->f); } end: @@ -1352,19 +1433,30 @@ static av_cold int cfhd_close(AVCodecContext *avctx) ff_free_vlc(&s->vlc_9); ff_free_vlc(&s->vlc_18); } + if (s->i_frame.f && s->i_frame.f->data[0]) + ff_thread_release_buffer(avctx, &s->i_frame); + if (s->p_frame.f && s->p_frame.f->data[0]) + ff_thread_release_buffer(avctx, &s->p_frame); + + if (s->i_frame.f) + av_frame_free(&s->i_frame.f); + if (s->p_frame.f) + av_frame_free(&s->p_frame.f); return 0; } AVCodec ff_cfhd_decoder = { - .name = "cfhd", - .long_name = NULL_IF_CONFIG_SMALL("Cineform HD"), - .type = AVMEDIA_TYPE_VIDEO, - .id = AV_CODEC_ID_CFHD, - .priv_data_size = sizeof(CFHDContext), - .init = cfhd_init, - .close = cfhd_close, - .decode = cfhd_decode, - .capabilities = AV_CODEC_CAP_DR1, - .caps_internal = FF_CODEC_CAP_INIT_CLEANUP, + .name = "cfhd", + .long_name = NULL_IF_CONFIG_SMALL("Cineform HD"), + .type = AVMEDIA_TYPE_VIDEO, + .id = AV_CODEC_ID_CFHD, + .priv_data_size = sizeof(CFHDContext), + .init = cfhd_init, + .close = cfhd_close, + .decode = cfhd_decode, + .init_thread_copy = ONLY_IF_THREADS_ENABLED(cfhd_init), + .update_thread_context = ONLY_IF_THREADS_ENABLED(update_thread_context), + .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS, + .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP, }; diff --git a/libavcodec/cfhd.h b/libavcodec/cfhd.h index 047c0f2028..d7a2ffe0a7 100644 --- a/libavcodec/cfhd.h +++ b/libavcodec/cfhd.h @@ -29,6 +29,7 @@ #include "bytestream.h" #include "get_bits.h" #include "vlc.h" +#include "thread.h" #define VLC_BITS 9 #define SUBBAND_COUNT 17 @@ -63,7 +64,7 @@ typedef struct Plane { /* TODO: merge this into SubBand structure */ int16_t *subband[SUBBAND_COUNT]; - int16_t *l_h[10]; + int16_t *l_h[12]; SubBand band[DWT_LEVELS][4]; } Plane; @@ -76,6 +77,10 @@ typedef struct Peak { typedef struct CFHDContext { AVCodecContext *avctx; + ThreadFrame i_frame; + ThreadFrame p_frame; + ThreadFrame *connection; + ThreadFrame *picture; CFHD_RL_VLC_ELEM table_9_rl_vlc[2088]; VLC vlc_9; @@ -116,6 +121,7 @@ typedef struct CFHDContext { uint8_t prescale_shift[3]; Plane plane[4]; + Plane *buffers; Peak peak; } CFHDContext;