16-bit gguf version of https://huggingface.co/meta-llama/Llama-2-7b-chat-hf
For quantized versions, see https://huggingface.co/models?search=thebloke/llama-2-7b-chat
- Downloads last month
- 4
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for pcuenq/Llama-2-7b-chat-gguf
Base model
meta-llama/Llama-2-7b-chat-hf