Update backend/query_llm.py
Browse files- backend/query_llm.py +2 -2
backend/query_llm.py
CHANGED
|
@@ -106,8 +106,8 @@ def generate_hf(prompt: str, history: str, temperature: float = 0.9, max_new_tok
|
|
| 106 |
for response in stream:
|
| 107 |
print(response.token.text)
|
| 108 |
output += response.token.text
|
| 109 |
-
|
| 110 |
-
|
| 111 |
|
| 112 |
except Exception as e:
|
| 113 |
if "Too Many Requests" in str(e):
|
|
|
|
| 106 |
for response in stream:
|
| 107 |
print(response.token.text)
|
| 108 |
output += response.token.text
|
| 109 |
+
#return output
|
| 110 |
+
yield output
|
| 111 |
|
| 112 |
except Exception as e:
|
| 113 |
if "Too Many Requests" in str(e):
|