|
@@ -51,7 +51,8 @@ class OpenLLM(LLM):
|
|
|
) -> str:
|
|
|
params = {
|
|
|
"prompt": prompt,
|
|
|
- "llm_config": self.llm_kwargs
|
|
|
+ "llm_config": self.llm_kwargs,
|
|
|
+ "stop": stop,
|
|
|
}
|
|
|
|
|
|
headers = {"Content-Type": "application/json"}
|
|
@@ -65,11 +66,11 @@ class OpenLLM(LLM):
|
|
|
raise ValueError(f"OpenLLM HTTP {response.status_code} error: {response.text}")
|
|
|
|
|
|
json_response = response.json()
|
|
|
- completion = json_response["responses"][0]
|
|
|
+ completion = json_response["outputs"][0]['text']
|
|
|
completion = completion.lstrip(prompt)
|
|
|
|
|
|
- if stop is not None:
|
|
|
- completion = enforce_stop_tokens(completion, stop)
|
|
|
+ # if stop is not None:
|
|
|
+ # completion = enforce_stop_tokens(completion, stop)
|
|
|
|
|
|
return completion
|
|
|
|