2025-09-16 17:23:50,202 xinference.core.worker 1241 INFO [request daefb0a8-92de-11f0-9b01-0242ac11000b] Enter launch_builtin_model, args: , kwargs: model_uid=m3e-large-0,model_name=m3e-large,model_size_in_billions=None,model_format=,quantization=,model_engine=,model_type=embedding,n_gpu=auto,request_limits=None,peft_model_config=None,gpu_idx=None,download_hub=None,model_path=None,enable_virtual_env=None,virtual_env_packages=None,envs=None,xavier_config=None 2025-09-16 17:23:50,959 xinference.core.worker 1241 ERROR Failed to load model m3e-large-0 Traceback (most recent call last): File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/worker.py", line 1042, in launch_builtin_model model = await asyncio.to_thread( ^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/asyncio/threads.py", line 25, in to_thread return await loop.run_in_executor(None, func_call) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/concurrent/futures/thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/core.py", line 59, in create_model_instance return create_embedding_model_instance( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/core.py", line 335, in create_embedding_model_instance embedding_cls = check_engine_by_model_name_and_engine( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/embed_family.py", line 131, in check_engine_by_model_name_and_engine raise ValueError(f"Model {model_name} cannot be run on engine {model_engine}.") ValueError: Model m3e-large cannot be run on engine . 2025-09-16 17:23:50,979 xinference.core.worker 1241 ERROR [request daefb0a8-92de-11f0-9b01-0242ac11000b] Leave launch_builtin_model, error: Model m3e-large cannot be run on engine ., elapsed time: 0 s Traceback (most recent call last): File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/utils.py", line 93, in wrapped ret = await func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/worker.py", line 1042, in launch_builtin_model model = await asyncio.to_thread( ^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/asyncio/threads.py", line 25, in to_thread return await loop.run_in_executor(None, func_call) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/concurrent/futures/thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/core.py", line 59, in create_model_instance return create_embedding_model_instance( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/core.py", line 335, in create_embedding_model_instance embedding_cls = check_engine_by_model_name_and_engine( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/embed_family.py", line 131, in check_engine_by_model_name_and_engine raise ValueError(f"Model {model_name} cannot be run on engine {model_engine}.") ValueError: Model m3e-large cannot be run on engine . 2025-09-16 17:23:50,990 xinference.api.restful_api 1222 ERROR [address=0.0.0.0:61445, pid=1241] Model m3e-large cannot be run on engine . Traceback (most recent call last): File "/root/miniconda3/lib/python3.12/site-packages/xinference/api/restful_api.py", line 1229, in launch_model model_uid = await (await self._get_supervisor_ref()).launch_builtin_model( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/context.py", line 262, in send return self._process_result_message(result) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/context.py", line 111, in _process_result_message raise message.as_instanceof_cause() File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/pool.py", line 689, in send result = await self._run_coro(message.message_id, coro) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/pool.py", line 389, in _run_coro return await coro File "/root/miniconda3/lib/python3.12/site-packages/xoscar/api.py", line 418, in __on_receive__ return await super().__on_receive__(message) # type: ignore ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 564, in __on_receive__ raise ex File "xoscar/core.pyx", line 526, in xoscar.core._BaseActor.__on_receive__ async with self._lock: ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 527, in xoscar.core._BaseActor.__on_receive__ with debug_async_timeout('actor_lock_timeout', ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 532, in xoscar.core._BaseActor.__on_receive__ result = await result ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/supervisor.py", line 1280, in launch_builtin_model await _launch_model() ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/supervisor.py", line 1215, in _launch_model subpool_address = await _launch_one_model( ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/supervisor.py", line 1166, in _launch_one_model subpool_address = await worker_ref.launch_builtin_model( ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/context.py", line 262, in send return self._process_result_message(result) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/context.py", line 111, in _process_result_message raise message.as_instanceof_cause() ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/pool.py", line 689, in send result = await self._run_coro(message.message_id, coro) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xoscar/backends/pool.py", line 389, in _run_coro return await coro File "/root/miniconda3/lib/python3.12/site-packages/xoscar/api.py", line 418, in __on_receive__ return await super().__on_receive__(message) # type: ignore ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 564, in __on_receive__ raise ex File "xoscar/core.pyx", line 526, in xoscar.core._BaseActor.__on_receive__ async with self._lock: ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 527, in xoscar.core._BaseActor.__on_receive__ with debug_async_timeout('actor_lock_timeout', ^^^^^^^^^^^^^^^^^ File "xoscar/core.pyx", line 532, in xoscar.core._BaseActor.__on_receive__ result = await result ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/utils.py", line 93, in wrapped ret = await func(*args, **kwargs) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/core/worker.py", line 1042, in launch_builtin_model model = await asyncio.to_thread( ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/asyncio/threads.py", line 25, in to_thread return await loop.run_in_executor(None, func_call) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/concurrent/futures/thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/core.py", line 59, in create_model_instance return create_embedding_model_instance( ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/core.py", line 335, in create_embedding_model_instance embedding_cls = check_engine_by_model_name_and_engine( ^^^^^^^^^^^^^^^^^ File "/root/miniconda3/lib/python3.12/site-packages/xinference/model/embedding/embed_family.py", line 131, in check_engine_by_model_name_and_engine raise ValueError(f"Model {model_name} cannot be run on engine {model_engine}.") ^^^^^^^^^^^^^^^^^ ValueError: [address=0.0.0.0:61445, pid=1241] Model m3e-large cannot be run on engine .