update LLM for response generation
Browse files
app.py
CHANGED
@@ -790,7 +790,8 @@ def respond(
|
|
790 |
# Stream response
|
791 |
response = client.chat.completions.create(
|
792 |
messages=[{"role": "user", "content": prompt}],
|
793 |
-
model="
|
|
|
794 |
stream=True,
|
795 |
)
|
796 |
cumulative_response = "" # Keep track of the cumulative response
|
|
|
790 |
# Stream response
|
791 |
response = client.chat.completions.create(
|
792 |
messages=[{"role": "user", "content": prompt}],
|
793 |
+
model="llama-3.1-8b-instant",
|
794 |
+
# model="llama-3.3-70b-versatile",
|
795 |
stream=True,
|
796 |
)
|
797 |
cumulative_response = "" # Keep track of the cumulative response
|