The model is the best for coding.
2
1
#7 opened about 22 hours ago
by
AekDevDev

When running with a single GPU, I get an error saying the VRAM is insufficient. However, when using multiple GPUs on a single machine, there are many errors. My vllm version is 0.8.4.
1
#6 opened 1 day ago
by
hanson888

BitsAndBytes quantization inference error
#5 opened 1 day ago
by
chengfy

Some bug when using function call with vllm==0.8.4
2
#4 opened 2 days ago
by
waple

SimpleQA Scores Are WAY off
2
5
#3 opened 4 days ago
by
phil111
Need fp8 version for inerface
#2 opened 4 days ago
by
iwaitu

RuntimeError: CUDA error: device-side assert triggered
#1 opened 4 days ago
by
DsnTgr