diff mbox series

[FFmpeg-devel] lavfi/dnn_backend_openvino.c: Fix Memory Leak in execute_model_ov

Message ID 20210618162307.20141-1-shubhanshu.e01@gmail.com
State Accepted
Commit f73943d514802ecc521b92ebf2d1595526968fd5
Headers show
Series [FFmpeg-devel] lavfi/dnn_backend_openvino.c: Fix Memory Leak in execute_model_ov | expand

Checks

Context Check Description
andriy/x86_make success Make finished
andriy/x86_make_fate success Make fate finished
andriy/PPC64_make success Make finished
andriy/PPC64_make_fate success Make fate finished

Commit Message

Shubhanshu Saxena June 18, 2021, 4:23 p.m. UTC
In cases where the execution inside the function execute_model_ov fails,
push the RequestItem back to the request_queue before returning the error.
In case pushing back fails, release the allocated memory.

Signed-off-by: Shubhanshu Saxena <shubhanshu.e01@gmail.com>
---
 libavfilter/dnn/dnn_backend_openvino.c | 12 +++++++++---
 1 file changed, 9 insertions(+), 3 deletions(-)

Comments

Guo, Yejun July 4, 2021, 11:39 a.m. UTC | #1
> -----Original Message-----
> From: ffmpeg-devel <ffmpeg-devel-bounces@ffmpeg.org> On Behalf Of
> Shubhanshu Saxena
> Sent: 2021年6月19日 0:23
> To: ffmpeg-devel@ffmpeg.org
> Cc: Shubhanshu Saxena <shubhanshu.e01@gmail.com>
> Subject: [FFmpeg-devel] [PATCH] lavfi/dnn_backend_openvino.c: Fix
> Memory Leak in execute_model_ov
> 
> In cases where the execution inside the function execute_model_ov fails,
> push the RequestItem back to the request_queue before returning the error.
> In case pushing back fails, release the allocated memory.
> 
> Signed-off-by: Shubhanshu Saxena <shubhanshu.e01@gmail.com>
> ---
>  libavfilter/dnn/dnn_backend_openvino.c | 12 +++++++++---
>  1 file changed, 9 insertions(+), 3 deletions(-)
> 
> diff --git a/libavfilter/dnn/dnn_backend_openvino.c
> b/libavfilter/dnn/dnn_backend_openvino.c
> index 702c4fb9ee..29ec8f6a8f 100644
> --- a/libavfilter/dnn/dnn_backend_openvino.c
> +++ b/libavfilter/dnn/dnn_backend_openvino.c
> @@ -448,12 +448,12 @@ static DNNReturnType
> execute_model_ov(RequestItem *request, Queue *inferenceq)
>          status = ie_infer_set_completion_callback(request->infer_request,
> &request->callback);
>          if (status != OK) {
>              av_log(ctx, AV_LOG_ERROR, "Failed to set completion callback for
> inference\n");
> -            return DNN_ERROR;
> +            goto err;
>          }
>          status = ie_infer_request_infer_async(request->infer_request);
>          if (status != OK) {
>              av_log(ctx, AV_LOG_ERROR, "Failed to start async inference\n");
> -            return DNN_ERROR;
> +            goto err;
>          }
>          return DNN_SUCCESS;
>      } else {
> @@ -464,11 +464,17 @@ static DNNReturnType
> execute_model_ov(RequestItem *request, Queue *inferenceq)
>          status = ie_infer_request_infer(request->infer_request);
>          if (status != OK) {
>              av_log(ctx, AV_LOG_ERROR, "Failed to start synchronous model
> inference\n");
> -            return DNN_ERROR;
> +            goto err;
>          }
>          infer_completion_callback(request);
>          return (task->inference_done == task->inference_todo) ?
> DNN_SUCCESS : DNN_ERROR;
>      }
> +err:
> +    if (ff_safe_queue_push_back(ov_model->request_queue, request) < 0)
> {
> +        ie_infer_request_free(&request->infer_request);
> +        av_freep(&request);
> +    }
> +    return DNN_ERROR;
>  }
> 
>  static DNNReturnType get_input_ov(void *model, DNNData *input, const
> char *input_name)
> --

LGTM, will push soon, thanks.
diff mbox series

Patch

diff --git a/libavfilter/dnn/dnn_backend_openvino.c b/libavfilter/dnn/dnn_backend_openvino.c
index 702c4fb9ee..29ec8f6a8f 100644
--- a/libavfilter/dnn/dnn_backend_openvino.c
+++ b/libavfilter/dnn/dnn_backend_openvino.c
@@ -448,12 +448,12 @@  static DNNReturnType execute_model_ov(RequestItem *request, Queue *inferenceq)
         status = ie_infer_set_completion_callback(request->infer_request, &request->callback);
         if (status != OK) {
             av_log(ctx, AV_LOG_ERROR, "Failed to set completion callback for inference\n");
-            return DNN_ERROR;
+            goto err;
         }
         status = ie_infer_request_infer_async(request->infer_request);
         if (status != OK) {
             av_log(ctx, AV_LOG_ERROR, "Failed to start async inference\n");
-            return DNN_ERROR;
+            goto err;
         }
         return DNN_SUCCESS;
     } else {
@@ -464,11 +464,17 @@  static DNNReturnType execute_model_ov(RequestItem *request, Queue *inferenceq)
         status = ie_infer_request_infer(request->infer_request);
         if (status != OK) {
             av_log(ctx, AV_LOG_ERROR, "Failed to start synchronous model inference\n");
-            return DNN_ERROR;
+            goto err;
         }
         infer_completion_callback(request);
         return (task->inference_done == task->inference_todo) ? DNN_SUCCESS : DNN_ERROR;
     }
+err:
+    if (ff_safe_queue_push_back(ov_model->request_queue, request) < 0) {
+        ie_infer_request_free(&request->infer_request);
+        av_freep(&request);
+    }
+    return DNN_ERROR;
 }
 
 static DNNReturnType get_input_ov(void *model, DNNData *input, const char *input_name)