From 57b51784131ec759b6672f8e6e948c8b7a3e3b23 Mon Sep 17 00:00:00 2001 From: "Mr.V" Date: Mon, 12 Aug 2024 09:44:49 +0800 Subject: [PATCH] =?UTF-8?q?1=E3=80=81=E4=BC=98=E5=8C=96=E8=AF=86=E5=88=AB?= =?UTF-8?q?=E5=88=87=E5=88=86=E6=95=88=E6=9E=9C=EF=BC=8C=E5=A2=9E=E5=8A=A0?= =?UTF-8?q?=E6=8A=A5=E8=AD=A6=E4=BF=A1=E6=81=AF?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- nvidia_ascend_tools/nvidia_tools/yolov5/src/inference.cu | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/nvidia_ascend_tools/nvidia_tools/yolov5/src/inference.cu b/nvidia_ascend_tools/nvidia_tools/yolov5/src/inference.cu index e894fb8..6db3eea 100644 --- a/nvidia_ascend_tools/nvidia_tools/yolov5/src/inference.cu +++ b/nvidia_ascend_tools/nvidia_tools/yolov5/src/inference.cu @@ -113,8 +113,8 @@ void Inference::doInference(IExecutionContext& context, cudaStream_t& stream, vo void Inference::doInferenceV2(IExecutionContext& context, cudaStream_t& stream, void **buffers, unsigned int outputIndex, float* output, int outputSize, int batchSize) { - context.enqueue(batchSize, buffers, stream, nullptr); - //context.enqueueV2(buffers, stream, nullptr); +// context.enqueue(batchSize, buffers, stream, nullptr); + context.enqueueV2(buffers, stream, nullptr); CUDA_CHECK(cudaMemcpyAsync(output, buffers[outputIndex], batchSize * outputSize * sizeof(float), cudaMemcpyDeviceToHost, stream)); cudaStreamSynchronize(stream); }