瀏覽代碼

fix: xinference chat completion error (#952)

takatost 1 年之前
父節點
當前提交
4f3053a8cc
共有 1 個文件被更改,包括 2 次插入2 次删除
  1. 2 2
      api/core/third_party/langchain/llms/xinference_llm.py

+ 2 - 2
api/core/third_party/langchain/llms/xinference_llm.py

@@ -46,7 +46,7 @@ class XinferenceLLM(Xinference):
                 return combined_text_output
                 return combined_text_output
             else:
             else:
                 completion = model.chat(prompt=prompt, generate_config=generate_config)
                 completion = model.chat(prompt=prompt, generate_config=generate_config)
-                return completion["choices"][0]["text"]
+                return completion["choices"][0]["message"]["content"]
         elif isinstance(model, RESTfulGenerateModelHandle):
         elif isinstance(model, RESTfulGenerateModelHandle):
             generate_config: "LlamaCppGenerateConfig" = kwargs.get("generate_config", {})
             generate_config: "LlamaCppGenerateConfig" = kwargs.get("generate_config", {})
 
 
@@ -82,7 +82,7 @@ class XinferenceLLM(Xinference):
                 completion = combined_text_output
                 completion = combined_text_output
             else:
             else:
                 completion = model.chat(prompt=prompt, generate_config=generate_config)
                 completion = model.chat(prompt=prompt, generate_config=generate_config)
-                completion = completion["choices"][0]["text"]
+                completion = completion["choices"][0]["message"]["content"]
 
 
             if stop is not None:
             if stop is not None:
                 completion = enforce_stop_tokens(completion, stop)
                 completion = enforce_stop_tokens(completion, stop)