Forráskód Böngészése

feat: support chatglm_turbo model #1443 (#1460)

takatost 1 éve
szülő
commit
4dfbcd0b4e

+ 5 - 0
api/core/model_providers/providers/zhipuai_provider.py

@@ -26,6 +26,11 @@ class ZhipuAIProvider(BaseModelProvider):
     def _get_fixed_model_list(self, model_type: ModelType) -> list[dict]:
         if model_type == ModelType.TEXT_GENERATION:
             return [
+                {
+                    'id': 'chatglm_turbo',
+                    'name': 'chatglm_turbo',
+                    'mode': ModelMode.CHAT.value,
+                },
                 {
                     'id': 'chatglm_pro',
                     'name': 'chatglm_pro',

+ 6 - 0
api/core/model_providers/rules/zhipuai.json

@@ -11,6 +11,12 @@
     },
     "model_flexibility": "fixed",
     "price_config": {
+        "chatglm_turbo": {
+            "prompt": "0.005",
+            "completion": "0.005",
+            "unit": "0.001",
+            "currency": "RMB"
+        },
         "chatglm_pro": {
             "prompt": "0.01",
             "completion": "0.01",

+ 1 - 1
api/core/third_party/langchain/llms/zhipuai_llm.py

@@ -96,7 +96,7 @@ class ZhipuAIChatLLM(BaseChatModel):
         return True
 
     client: Any = None  #: :meta private:
-    model: str = "chatglm_lite"
+    model: str = "chatglm_turbo"
     """Model name to use."""
     temperature: float = 0.95
     """A non-negative float that tunes the degree of randomness in generation."""