this post was submitted on 10 Nov 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[โ€“] DaniyarQQQ@alien.top 1 points 10 months ago (1 children)

Looks like it can work with AWQ models. Can it work with GPTQ (Exllama2) and GGUF models?

[โ€“] davidmezzetti@alien.top 1 points 10 months ago

It works with GPTQ models as well, just need to install AutoGPTQ.

You would need to replace the LLM pipeline with llama.cpp for it to work with GGUF models.

See this page for more: https://huggingface.co/docs/transformers/main_classes/quantization