Spaces:
Running
on
Zero
Running
on
Zero
jenbenarye
commited on
Commit
·
afda8d0
1
Parent(s):
a287222
changed model temp
Browse files- app/app.py +14 -3
app/app.py
CHANGED
@@ -159,6 +159,9 @@ def add_user_message(history, message):
|
|
159 |
|
160 |
|
161 |
def format_system_message(language: str):
|
|
|
|
|
|
|
162 |
system_message = [
|
163 |
{
|
164 |
"role": "system",
|
@@ -169,7 +172,8 @@ def format_system_message(language: str):
|
|
169 |
"content": f"Start by asking me a question in {language}."
|
170 |
}
|
171 |
]
|
172 |
-
response = call_pipeline(system_message, temperature=
|
|
|
173 |
new_system_message = [
|
174 |
{
|
175 |
"role": "system",
|
@@ -285,6 +289,8 @@ def add_fake_like_data(
|
|
285 |
@spaces.GPU
|
286 |
def call_pipeline(messages: list, temperature: float = 0.7):
|
287 |
"""Call the appropriate model pipeline based on configuration"""
|
|
|
|
|
288 |
if ZERO_GPU:
|
289 |
tokenizer = CLIENT["tokenizer"]
|
290 |
# Ensure messages follow the proper alternating pattern
|
@@ -330,6 +336,7 @@ def call_pipeline(messages: list, temperature: float = 0.7):
|
|
330 |
return_full_text=False,
|
331 |
temperature=temperature,
|
332 |
do_sample=True,
|
|
|
333 |
)
|
334 |
|
335 |
return response[0]["generated_text"]
|
@@ -522,12 +529,16 @@ def wrangle_retry_data(
|
|
522 |
language=language,
|
523 |
)
|
524 |
|
|
|
|
|
|
|
|
|
525 |
# Return the history without a new message
|
526 |
history = respond(
|
527 |
history=history[:-1],
|
528 |
language=language,
|
529 |
-
temperature=
|
530 |
-
seed=
|
531 |
)
|
532 |
return history, update_dataframe(dataframe, history)
|
533 |
|
|
|
159 |
|
160 |
|
161 |
def format_system_message(language: str):
|
162 |
+
# Use a higher temperature with randomization for more diversity
|
163 |
+
random_temp = random.uniform(1.3, 2.0) # More random between 1.3 and 2.0
|
164 |
+
|
165 |
system_message = [
|
166 |
{
|
167 |
"role": "system",
|
|
|
172 |
"content": f"Start by asking me a question in {language}."
|
173 |
}
|
174 |
]
|
175 |
+
response = call_pipeline(system_message, temperature=random_temp)
|
176 |
+
|
177 |
new_system_message = [
|
178 |
{
|
179 |
"role": "system",
|
|
|
289 |
@spaces.GPU
|
290 |
def call_pipeline(messages: list, temperature: float = 0.7):
|
291 |
"""Call the appropriate model pipeline based on configuration"""
|
292 |
+
|
293 |
+
|
294 |
if ZERO_GPU:
|
295 |
tokenizer = CLIENT["tokenizer"]
|
296 |
# Ensure messages follow the proper alternating pattern
|
|
|
336 |
return_full_text=False,
|
337 |
temperature=temperature,
|
338 |
do_sample=True,
|
339 |
+
top_p=0.9, # Add top_p sampling for more diversity
|
340 |
)
|
341 |
|
342 |
return response[0]["generated_text"]
|
|
|
529 |
language=language,
|
530 |
)
|
531 |
|
532 |
+
# Use randomized temperature for more varied responses when retrying
|
533 |
+
random_temp = random.randint(70, 150) / 100 # Between 0.7 and 1.5
|
534 |
+
random_seed = random.randint(0, 1000000)
|
535 |
+
|
536 |
# Return the history without a new message
|
537 |
history = respond(
|
538 |
history=history[:-1],
|
539 |
language=language,
|
540 |
+
temperature=random_temp,
|
541 |
+
seed=random_seed,
|
542 |
)
|
543 |
return history, update_dataframe(dataframe, history)
|
544 |
|