Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF woes π΅βπ« #16253
-
I wanted to run https://huggingface.co/ggml-org/Qwen3-30B-A3B-Instruct-2507-Q8_0-GGUF on a MacBook Pro M3 with 36GB and noticed that Since the model was built with gguf-my-repo I tried to create my own But when I load it with:
I get this log.txt What am I doing wrong? How can I fix it? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
You are running out of memory because the context of this model is 256k and requires ~25GB. You should be able to run with |
Beta Was this translation helpful? Give feedback.
You are running out of memory because the context of this model is 256k and requires ~25GB. You should be able to run with
-c 32768
and probably higher depending on how much free memory you have.