│ /usr/lib/python3.11/concurrent/futures/thread.p │[Nest] 741 - 02/26/2026, 4:48:23 PM ERROR [Api:ErrorInterceptor~5ipss87g] Unknown error: Error: Machine learning request '{"clip":{"textual":{"modelName":"XLM-Roberta-Large-Vit-B-32","options":{"language":"nl-NL"}}}}' failed for all URLs Error: Machine learning request '{"clip":{"textual":{"modelName":"XLM-Roberta-Large-Vit-B-32","options":{"language":"nl-NL"}}}}' failed for all URLs at MachineLearningRepository.predict (/app/immich/server/dist/repositories/machine-learning.repository.js:117:15) │ y:58 in run │ at process.processTicksAndRejections (node:internal/process/task_queues:103:5) at async MachineLearningRepository.encodeText (/app/immich/server/dist/repositories/machine-learning.repository.js:140:26) at async SearchService.searchSmart (/app/immich/server/dist/services/search.service.js:94:29) │ │ │ /app/immich/machine-learning/immich_ml/main.py: │ │ 241 in _load │ │ │ │ 238 │ │ │ raise HTTPException(500, f"Fa │ │ 239 │ │ with lock: │ │ 240 │ │ │ try: │ │ ❱ 241 │ │ │ │ model.load() │ │ 242 │ │ │ except FileNotFoundError as e │ │ 243 │ │ │ │ if model.model_format == │ │ 244 │ │ │ │ │ raise e │ │ │ │ /app/immich/machine-learning/immich_ml/models/b │ │ ase.py:53 in load │ │ │ │ 50 │ │ self.download() │ │ 51 │ │ attempt = f"Attempt #{self.load_a │ │ else "Loading" │ │ 52 │ │ log.info(f"{attempt} {self.model_ │ │ '{self.model_name}' to memory") │ │ ❱ 53 │ │ self.session = self._load() │ │ 54 │ │ self.loaded = True │ │ 55 │ │ │ 56 │ def predict(self, *inputs: Any, **mod │ │ │ │ /app/immich/machine-learning/immich_ml/models/c │ │ lip/textual.py:28 in _load │ │ │ │ 25 │ │ return serialize_np_array(res) │ │ 26 │ │ │ 27 │ def _load(self) -> ModelSession: │ │ ❱ 28 │ │ session = super()._load() │ │ 29 │ │ log.debug(f"Loading tokenizer for │ │ 30 │ │ self.tokenizer = self._load_token │ │ 31 │ │ tokenizer_kwargs: dict[str, Any] │ │ │ │ /app/immich/machine-learning/immich_ml/models/b │ │ ase.py:83 in _load │ │ │ │ 80 │ │ ) │ │ 81 │ │ │ 82 │ def _load(self) -> ModelSession: │ │ ❱ 83 │ │ return self._make_session(self.mo │ │ 84 │ │ │ 85 │ def clear_cache(self) -> None: │ │ 86 │ │ if not self.cache_dir.exists(): │ │ │ │ /app/immich/machine-learning/immich_ml/models/b │ │ ase.py:115 in _make_session │ │ │ │ 112 │ │ │ case ".armnn": │ │ 113 │ │ │ │ session: ModelSession = A │ │ 114 │ │ │ case ".onnx": │ │ ❱ 115 │ │ │ │ session = OrtSession(mode │ │ 116 │ │ │ case ".rknn": │ │ 117 │ │ │ │ session = rknn.RknnSessio │ │ 118 │ │ │ case _: │ │ │ │ /app/immich/machine-learning/immich_ml/sessions │ │ /ort.py:30 in __init__ │ │ │ │ 27 │ │ self.providers = providers if pro │ │ 28 │ │ self.provider_options = provider_ │ │ self._provider_options_default │ │ 29 │ │ self.sess_options = sess_options │ │ self._sess_options_default │ │ ❱ 30 │ │ self.session = ort.InferenceSessi │ │ 31 │ │ │ self.model_path.as_posix(), │ │ 32 │ │ │ providers=self.providers, │ │ 33 │ │ │ provider_options=self.provide │ │ │ │ /lsiopy/lib/python3.11/site-packages/onnxruntim │ │ e/capi/onnxruntime_inference_collection.py:485 │ │ in __init__ │ │ │ │ 482 │ │ disabled_optimizers = kwargs.get │ │ 483 │ │ │ │ 484 │ │ try: │ │ ❱ 485 │ │ │ self._create_inference_sessi │ │ disabled_optimizers) │ │ 486 │ │ except (ValueError, RuntimeError │ │ 487 │ │ │ if self._enable_fallback: │ │ 488 │ │ │ │ try: │ │ │ │ /lsiopy/lib/python3.11/site-packages/onnxruntim │ │ e/capi/onnxruntime_inference_collection.py:584 │ │ in _create_inference_session │ │ │ │ 581 │ │ │ disabled_optimizers = set(di │ │ 582 │ │ │ │ 583 │ │ # initialize the C++ InferenceSe │ │ ❱ 584 │ │ sess.initialize_session(provider │ │ 585 │ │ │ │ 586 │ │ self._sess = sess │ │ 587 │ │ self._sess_options = self._sess. │ ╰─────────────────────────────────────────────────╯ RuntimeException: [ONNXRuntimeError] : 6 : RUNTIME_EXCEPTION : Exception during initialization: /onnxruntime_src/onnxruntime/core/framework/bfc_are na.cc:359 void* onnxruntime::BFCArena::AllocateRawInternal(size_t, bool, onnxruntime::Stream*) Failed to allocate memory for requested buffer of size 1024008192