jenbenarye commited on
Commit
afda8d0
·
1 Parent(s): a287222

changed model temp

Browse files
Files changed (1) hide show
  1. app/app.py +14 -3
app/app.py CHANGED
@@ -159,6 +159,9 @@ def add_user_message(history, message):
159
 
160
 
161
  def format_system_message(language: str):
 
 
 
162
  system_message = [
163
  {
164
  "role": "system",
@@ -169,7 +172,8 @@ def format_system_message(language: str):
169
  "content": f"Start by asking me a question in {language}."
170
  }
171
  ]
172
- response = call_pipeline(system_message, temperature=1.0)
 
173
  new_system_message = [
174
  {
175
  "role": "system",
@@ -285,6 +289,8 @@ def add_fake_like_data(
285
  @spaces.GPU
286
  def call_pipeline(messages: list, temperature: float = 0.7):
287
  """Call the appropriate model pipeline based on configuration"""
 
 
288
  if ZERO_GPU:
289
  tokenizer = CLIENT["tokenizer"]
290
  # Ensure messages follow the proper alternating pattern
@@ -330,6 +336,7 @@ def call_pipeline(messages: list, temperature: float = 0.7):
330
  return_full_text=False,
331
  temperature=temperature,
332
  do_sample=True,
 
333
  )
334
 
335
  return response[0]["generated_text"]
@@ -522,12 +529,16 @@ def wrangle_retry_data(
522
  language=language,
523
  )
524
 
 
 
 
 
525
  # Return the history without a new message
526
  history = respond(
527
  history=history[:-1],
528
  language=language,
529
- temperature=random.randint(1, 100) / 100,
530
- seed=random.randint(0, 1000000),
531
  )
532
  return history, update_dataframe(dataframe, history)
533
 
 
159
 
160
 
161
  def format_system_message(language: str):
162
+ # Use a higher temperature with randomization for more diversity
163
+ random_temp = random.uniform(1.3, 2.0) # More random between 1.3 and 2.0
164
+
165
  system_message = [
166
  {
167
  "role": "system",
 
172
  "content": f"Start by asking me a question in {language}."
173
  }
174
  ]
175
+ response = call_pipeline(system_message, temperature=random_temp)
176
+
177
  new_system_message = [
178
  {
179
  "role": "system",
 
289
  @spaces.GPU
290
  def call_pipeline(messages: list, temperature: float = 0.7):
291
  """Call the appropriate model pipeline based on configuration"""
292
+
293
+
294
  if ZERO_GPU:
295
  tokenizer = CLIENT["tokenizer"]
296
  # Ensure messages follow the proper alternating pattern
 
336
  return_full_text=False,
337
  temperature=temperature,
338
  do_sample=True,
339
+ top_p=0.9, # Add top_p sampling for more diversity
340
  )
341
 
342
  return response[0]["generated_text"]
 
529
  language=language,
530
  )
531
 
532
+ # Use randomized temperature for more varied responses when retrying
533
+ random_temp = random.randint(70, 150) / 100 # Between 0.7 and 1.5
534
+ random_seed = random.randint(0, 1000000)
535
+
536
  # Return the history without a new message
537
  history = respond(
538
  history=history[:-1],
539
  language=language,
540
+ temperature=random_temp,
541
+ seed=random_seed,
542
  )
543
  return history, update_dataframe(dataframe, history)
544