Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
Falcon-7B-Instruct-GPTQ

Text Generation
Transformers
Safetensors
English
RefinedWebModel
custom_code
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions Community
19
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: float key.dtype: float and value.dtype: c10::Half instead.

๐Ÿ‘ 2
#19 opened over 1 year ago by
zayuki

The model 'RWGPTQForCausalLM' is not supported for text-generation.

#18 opened over 1 year ago by
herMaster

Model not working for CPU

#17 opened over 1 year ago by
vivek0797

ValueError: Unrecognized configuration class

1
#14 opened almost 2 years ago by
hfgdfdsd

Can't use with tgi. Getting `RuntimeError: weight transformer.h.0.self_attention.query_key_value.weight does not exist`

1
#12 opened almost 2 years ago by
mpronesti

Integration to transformers pipeline

5
#10 opened almost 2 years ago by
clementdesroches

Custom 4-bit Finetuning 5-7 times faster inference than QLora

๐Ÿค 1
2
#5 opened almost 2 years ago by
rmihaylov

Getting 0 tokens while running using text-generation -webui

6
#4 opened almost 2 years ago by
avatar8875

CUDA extension not installed

3
#3 opened almost 2 years ago by
kllisre

Do you know anything about this error?

5
#2 opened almost 2 years ago by
RedXeol
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs