mmcgovern574 commited on
Commit
2efcfde
·
verified ·
1 Parent(s): f693daa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -2
app.py CHANGED
@@ -129,8 +129,7 @@ try:
129
  MODEL_ID,
130
  device_map="auto",
131
  quantization_config=quantization_config,
132
- torch_dtype=torch.bfloat16,
133
- attn_implementation="flash_attention_2" # Enable Flash Attention 2
134
  )
135
 
136
  # Apply Better Transformer optimization
 
129
  MODEL_ID,
130
  device_map="auto",
131
  quantization_config=quantization_config,
132
+ torch_dtype=torch.bfloat16
 
133
  )
134
 
135
  # Apply Better Transformer optimization