From c5d7e088482102ff439e1004b8418603449b6523 Mon Sep 17 00:00:00 2001 From: liunux4odoo Date: Wed, 15 Jan 2025 17:13:37 +0800 Subject: [PATCH] remove duplicated conversion; distincwish vision model --- xinference/model/llm/utils.py | 3 ++- xinference/model/llm/vllm/core.py | 1 - 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/xinference/model/llm/utils.py b/xinference/model/llm/utils.py index f6525d78a8..c5b861aa32 100644 --- a/xinference/model/llm/utils.py +++ b/xinference/model/llm/utils.py @@ -104,7 +104,8 @@ def get_full_context( tokenize=False, **kwargs, ): - messages = self.convert_messages_with_content_list_to_str_conversion(messages) + if "vision" not in self.model_family.model_ability: + messages = self.convert_messages_with_content_list_to_str_conversion(messages) if tokenizer is not None: try: full_context = tokenizer.apply_chat_template( diff --git a/xinference/model/llm/vllm/core.py b/xinference/model/llm/vllm/core.py index a2a1ece6c0..b9a8e5388b 100644 --- a/xinference/model/llm/vllm/core.py +++ b/xinference/model/llm/vllm/core.py @@ -804,7 +804,6 @@ async def async_chat( generate_config: Optional[Dict] = None, request_id: Optional[str] = None, ) -> Union[ChatCompletion, AsyncGenerator[ChatCompletionChunk, None]]: - messages = self.convert_messages_with_content_list_to_str_conversion(messages) tools = generate_config.pop("tools", []) if generate_config else None model_family = self.model_family.model_family or self.model_family.model_name full_context_kwargs = {}