OOM on 12 GB loading the model

#1
by mapa17 - opened

Hello,

i am running out of memory when loading the model with 12GB VRAM. What is the minimum memory needed for an AWQ quantized 7B model with group size of 128?

thx,
Manuel

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment