Qwen-7B-Chat-GGUF
์ด ์ ์ฅ์๋ Qwen/Qwen-7B-Chat ๋ชจ๋ธ์ GGUF๋ก ๋ณํํ๊ณ ์์ํํ ๊ฒฐ๊ณผ๋ฌผ์
๋๋ค.
์์ฑ ์ ๋ณด
- ์๋ณธ ๋ชจ๋ธ:
Qwen/Qwen-7B-Chat - ์๋ณธ revision:
main - ์ค๋ ๋ณํฉ ์ฌ์ฉ:
False - ๊ธฐ๋ณธ GGUF ํ์
:
f16 - ์์ํ ํ์
:
Q2_K,Q3_K,Q4_K,Q5_K,Q6_K,Q8_0
ํ์ผ ๋ชฉ๋ก
Qwen-7B-Chat-f16.ggufQwen-7B-Chat-Q2_K.ggufQwen-7B-Chat-Q3_K.ggufQwen-7B-Chat-Q4_K.ggufQwen-7B-Chat-Q5_K.ggufQwen-7B-Chat-Q6_K.ggufQwen-7B-Chat-Q8_0.ggufquantizer-manifest.json
์ฌ์ฉ ์์
./llama-cli -m Qwen-7B-Chat-f16.gguf -p "์๋
ํ์ธ์"
์ฐธ๊ณ
- ๋ณํ/์์ํ๋
llama.cpp๋๊ตฌ ์ฒด์ธ์ ์ฌ์ฉํ์ต๋๋ค. - ์์ฑ ๊ณผ์ ๋ฉํ๋ฐ์ดํฐ๋
quantizer-manifest.jsonํ์ผ์ ํฌํจ๋์ด ์์ต๋๋ค.
- Downloads last month
- 264
Hardware compatibility
Log In to add your hardware
2-bit
6-bit
8-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for h31337/Qwen-7B-Chat-GGUF
Base model
Qwen/Qwen-7B-Chat