Parcourir la source

Fix/4742 ollama num gpu option not consistent with allowed values (#4751)

crazywoola il y a 10 mois
Parent
commit
705a6e3a8e
1 fichiers modifiés avec 7 ajouts et 5 suppressions
  1. 7 5
      api/core/model_runtime/model_providers/ollama/llm/llm.py

+ 7 - 5
api/core/model_runtime/model_providers/ollama/llm/llm.py

@@ -534,12 +534,14 @@ class OllamaLargeLanguageModel(LargeLanguageModel):
                 ),
                 ParameterRule(
                     name='num_gpu',
-                    label=I18nObject(en_US="Num GPU"),
+                    label=I18nObject(en_US="GPU Layers"),
                     type=ParameterType.INT,
-                    help=I18nObject(en_US="The number of layers to send to the GPU(s). "
-                                          "On macOS it defaults to 1 to enable metal support, 0 to disable."),
-                    min=0,
-                    max=1
+                    help=I18nObject(en_US="The number of layers to offload to the GPU(s). "
+                                          "On macOS it defaults to 1 to enable metal support, 0 to disable."
+                                          "As long as a model fits into one gpu it stays in one. "
+                                          "It does not set the number of GPU(s). "),
+                    min=-1,
+                    default=1
                 ),
                 ParameterRule(
                     name='num_thread',