│ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); MIOpen Error: /long_pathname_so_that_rpms_can_package_the_debug_info/src/MLOpen/src/ocl/convolutionocl.cpp:869: Buffers cannot be NULL 2025-10-30 19:40:32.776511538 [E:onnxruntime:Default, rocm_call.cc:119 RocmCall] MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); 2025-10-30 19:40:32.776526479 [E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); [10/30/25 19:40:32] ERROR Exception in ASGI application ╭─────── Traceback (most recent call last) ───────╮ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:90 in │ │ __call__ │ │ │ │ 87 │ def __call__(self, input_content: np. │ │ 88 │ │ input_dict = dict(zip(self.get_in │ │ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); MIOpen(HIP): Warning [IsEnoughWorkspace] [GetSolutionsFallback AI] Solver , workspace required: 41287680, provided ptr: 0x7f2c269d8000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [EvaluateInvokers] Solver , workspace required: 41287680, provided ptr: 0x7f2c269d8000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [GetSolutionsFallback AI] Solver , workspace required: 46694400, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [EvaluateInvokers] Solver , workspace required: 46694400, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [GetSolutionsFallback AI] Solver , workspace required: 41287680, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [EvaluateInvokers] Solver , workspace required: 41287680, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [GetSolutionsFallback AI] Solver , workspace required: 41287680, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [EvaluateInvokers] Solver , workspace required: 41287680, provided ptr: 0x7f2c23200000 size: 33554432 MIOpen Error: /long_pathname_so_that_rpms_can_package_the_debug_info/src/MLOpen/src/ocl/convolutionocl.cpp:869: Buffers cannot be NULL 2025-10-30 19:40:35.925167895 [E:onnxruntime:Default, rocm_call.cc:119 RocmCall] MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); 2025-10-30 19:40:35.925176315 [E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); [10/30/25 19:40:35] ERROR Exception in ASGI application ╭─────── Traceback (most recent call last) ───────╮ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:90 in │ │ __call__ │ │ │ │ 87 │ def __call__(self, input_content: np. │ │ 88 │ │ input_dict = dict(zip(self.get_in │ │ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); MIOpen(HIP): Warning [IsEnoughWorkspace] [GetSolutionsFallback AI] Solver , workspace required: 46694400, provided ptr: 0x7f2c26000000 size: 33554432 MIOpen(HIP): Warning [IsEnoughWorkspace] [EvaluateInvokers] Solver , workspace required: 46694400, provided ptr: 0x7f2c26000000 size: 33554432 MIOpen Error: /long_pathname_so_that_rpms_can_package_the_debug_info/src/MLOpen/src/ocl/convolutionocl.cpp:869: Buffers cannot be NULL 2025-10-30 19:40:37.474975228 [E:onnxruntime:Default, rocm_call.cc:119 RocmCall] MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); 2025-10-30 19:40:37.474984438 [E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); [10/30/25 19:40:37] ERROR Exception in ASGI application ╭─────── Traceback (most recent call last) ───────╮ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:90 in │ │ __call__ │ │ │ │ 87 │ def __call__(self, input_content: np. │ │ 88 │ │ input_dict = dict(zip(self.get_in │ │ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); MIOpen Error: /long_pathname_so_that_rpms_can_package_the_debug_info/src/MLOpen/src/ocl/convolutionocl.cpp:869: Buffers cannot be NULL 2025-10-30 19:40:37.627549348 [E:onnxruntime:Default, rocm_call.cc:119 RocmCall] MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); 2025-10-30 19:40:37.627558618 [E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); [10/30/25 19:40:37] ERROR Exception in ASGI application ╭─────── Traceback (most recent call last) ───────╮ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:90 in │ │ __call__ │ │ │ │ 87 │ def __call__(self, input_content: np. │ │ 88 │ │ input_dict = dict(zip(self.get_in │ │ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); MIOpen Error: /long_pathname_so_that_rpms_can_package_the_debug_info/src/MLOpen/src/ocl/convolutionocl.cpp:869: Buffers cannot be NULL 2025-10-30 19:40:38.219914114 [E:onnxruntime:Default, rocm_call.cc:119 RocmCall] MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); 2025-10-30 19:40:38.219926184 [E:onnxruntime:, sequential_executor.cc:516 ExecuteKernel] Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/rocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); [10/30/25 19:40:38] ERROR Exception in ASGI application ╭─────── Traceback (most recent call last) ───────╮ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:90 in │ │ __call__ │ │ │ │ 87 │ def __call__(self, input_content: np. │ │ 88 │ │ input_dict = dict(zip(self.get_in │ │ 89 │ │ try: │ │ ❱ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ │ │ /opt/venv/lib/python3.10/site-packages/onnxrunt │ │ ime/capi/onnxruntime_inference_collection.py:26 │ │ 6 in run │ │ │ │ 263 │ │ if not output_names: │ │ 264 │ │ │ output_names = [output.name │ │ 265 │ │ try: │ │ ❱ 266 │ │ │ return self._sess.run(output │ │ 267 │ │ except C.EPFail as err: │ │ 268 │ │ │ if self._enable_fallback: │ │ 269 │ │ │ │ print(f"EP Error: {err!s │ ╰─────────────────────────────────────────────────╯ Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false); The above exception was the direct cause of the following exception: ╭─────── Traceback (most recent call last) ───────╮ │ /usr/src/immich_ml/main.py:177 in predict │ │ │ │ 174 │ │ inputs = text │ │ 175 │ else: │ │ 176 │ │ raise HTTPException(400, "Either │ │ ❱ 177 │ response = await run_inference(inputs │ │ 178 │ return ORJSONResponse(response) │ │ 179 │ │ 180 │ │ │ │ /usr/src/immich_ml/main.py:202 in run_inference │ │ │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ 201 │ without_deps, with_deps = entries │ │ ❱ 202 │ await asyncio.gather(*[_run_inference │ │ 203 │ if with_deps: │ │ 204 │ │ await asyncio.gather(*[_run_infer │ │ 205 │ if isinstance(payload, Image): │ │ │ │ /usr/src/immich_ml/main.py:197 in │ │ _run_inference │ │ │ │ 194 │ │ │ │ message = f"Task {entry[' │ │ output of {dep}" │ │ 195 │ │ │ │ raise HTTPException(400, │ │ 196 │ │ model = await load(model) │ │ ❱ 197 │ │ output = await run(model.predict, │ │ 198 │ │ outputs[model.identity] = output │ │ 199 │ │ response[entry["task"]] = output │ │ 200 │ │ │ │ /usr/src/immich_ml/main.py:215 in run │ │ │ │ 212 │ if thread_pool is None: │ │ 213 │ │ return func(*args, **kwargs) │ │ 214 │ partial_func = partial(func, *args, * │ │ ❱ 215 │ return await asyncio.get_running_loop │ │ 216 │ │ 217 │ │ 218 async def load(model: InferenceModel) -> │ │ │ │ /usr/lib/python3.10/concurrent/futures/thread.p │ │ y:58 in run │ │ │ │ /usr/src/immich_ml/models/base.py:60 in predict │ │ │ │ 57 │ │ self.load() │ │ 58 │ │ if model_kwargs: │ │ 59 │ │ │ self.configure(**model_kwargs │ │ ❱ 60 │ │ return self._predict(*inputs) │ │ 61 │ │ │ 62 │ @abstractmethod │ │ 63 │ def _predict(self, *inputs: Any, **mo │ │ │ │ /usr/src/immich_ml/models/ocr/detection.py:68 │ │ in _predict │ │ │ │ 65 │ │ return session │ │ 66 │ │ │ 67 │ def _predict(self, inputs: bytes | Ima │ │ ❱ 68 │ │ results = self.model(decode_cv2(in │ │ 69 │ │ if results.boxes is None or result │ │ 70 │ │ │ return self._empty │ │ 71 │ │ return { │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /ch_ppocr_det/main.py:59 in __call__ │ │ │ │ 56 │ │ if prepro_img is None: │ │ 57 │ │ │ return TextDetOutput() │ │ 58 │ │ │ │ ❱ 59 │ │ preds = self.session(prepro_img) │ │ 60 │ │ boxes, scores = self.postprocess_ │ │ 61 │ │ if len(boxes) < 1: │ │ 62 │ │ │ return TextDetOutput() │ │ │ │ /opt/venv/lib/python3.10/site-packages/rapidocr │ │ /inference_engine/onnxruntime/main.py:93 in │ │ __call__ │ │ │ │ 90 │ │ │ return self.session.run(self. │ │ 91 │ │ except Exception as e: │ │ 92 │ │ │ error_info = traceback.format │ │ ❱ 93 │ │ │ raise ONNXRuntimeError(error_ │ │ 94 │ │ │ 95 │ def get_input_names(self) -> List[str │ │ 96 │ │ return [v.name for v in self.sess │ ╰─────────────────────────────────────────────────╯ ONNXRuntimeError: Traceback (most recent call last): File "/opt/venv/lib/python3.10/site-packages/rapidocr/in ference_engine/onnxruntime/main.py", line 90, in __call__ return self.session.run(self.get_output_names(), input_dict)[0] File "/opt/venv/lib/python3.10/site-packages/onnxruntime /capi/onnxruntime_inference_collection.py", line 266, in run return self._sess.run(output_names, input_feed, run_options) onnxruntime.capi.onnxruntime_pybind11_state.Fail: [ONNXRuntimeError] : 1 : FAIL : Non-zero status code returned while running ConvTranspose node. Name:'ConvTranspose.0' Status Message: MIOPEN failure 3: miopenStatusBadParm ; GPU=0 ; hostname=ea3d28c568a9 ; file=/code/onnxruntime/onnxruntime/core/providers/r ocm/nn/conv_transpose.cc ; line=133 ; expr=miopenFindConvolutionBackwardDataAlgorithm( GetMiopenHandle(context), s_.x_tensor, x_data, s_.w_desc, w_data, s_.conv_desc, s_.y_tensor, y_data, 1, &algo_count, &perf, algo_search_workspace.get(), AlgoSearchWorkspaceSize, false);