Added New Model & README.md file
Browse files- Note.txt +11 -0
- README.md +44 -0
- orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm +3 -0
- orpheus_3b_0.1_ft_w8a8_RK3588_GGUF_F16.rkllm +3 -0
Note.txt
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ret =llm16.export_rkllm(f"orpheus-3b-0.1-ft_{quantized_dtype}_{target_platform}_16bit.rkllm")
|
| 2 |
+
INFO: Setting chat_template to "<|start_header_id|>system<|end_header_id|>\n\nCutting Knowledge Date: December 2023\nToday Date: 04 May 2025\n\n<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n[content]<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
| 3 |
+
INFO: Setting token_id of bos to 128000
|
| 4 |
+
INFO: Setting token_id of eos to 128009
|
| 5 |
+
INFO: Setting token_id of pad to 128004
|
| 6 |
+
INFO: Setting add_bos_token to True
|
| 7 |
+
Converting model: 100%|ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 255/255 [00:00<00:00, 1331939.63it/s]
|
| 8 |
+
INFO: Setting max_context_limit to 4096
|
| 9 |
+
INFO: Exporting the model, please wait ....
|
| 10 |
+
[=================================================>] 597/597 (100%)
|
| 11 |
+
INFO: Model has been saved to orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm!
|
README.md
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
base_model:
|
| 6 |
+
- Prince-1/orpheus_3b_0.1_4bit
|
| 7 |
+
- Prince-1/orpheus_3b_0.1_GGUF
|
| 8 |
+
tags:
|
| 9 |
+
- rkllm
|
| 10 |
+
- text-to-speech
|
| 11 |
+
- tts
|
| 12 |
+
- transformers
|
| 13 |
+
- llama
|
| 14 |
+
---
|
| 15 |
+
|
| 16 |
+
# Orpheus_3b_0.1_rkllm
|
| 17 |
+
|
| 18 |
+
**Orpheus_3b_0.1_rkllm** is a [Text-to-Speech](https://huggingface.co/models?pipeline_tag=text-to-speech&sort=trending) model built from Orpheus [16bit](https://huggingface.co/Prince-1/orpheus_3b_0.1_ft_16bit) and [GGUF F16](https://huggingface.co/Prince-1/orpheus_3b_0.1_GGUF), using the [Rkllm-Toolkit](https://github.com/airockchip/rknn-llm).
|
| 19 |
+
|
| 20 |
+
## Features
|
| 21 |
+
|
| 22 |
+
- ποΈ Text-to-Speech capability with optimized inference
|
| 23 |
+
- π§ Built from Orpheus 16bit & GGUF F16 formats
|
| 24 |
+
- π Runs on **RK3588 NPU** using **w8a8 quantization**
|
| 25 |
+
- βοΈ Powered by [RKLLM Toolkit v1.2.1b1](https://github.com/airockchip/rknn-llm)
|
| 26 |
+
- β‘ Designed for high-performance on edge devices
|
| 27 |
+
|
| 28 |
+
## Requirements
|
| 29 |
+
|
| 30 |
+
- RK3588-based device
|
| 31 |
+
- RKLLM Toolkit v1.2.1b1
|
| 32 |
+
- Compatible runtime environment for deploying quantized models
|
| 33 |
+
|
| 34 |
+
## Usage
|
| 35 |
+
|
| 36 |
+
1. Clone or download the model from [Hugging Face](https://huggingface.co/Prince-1/orpheus_3b_0.1_rkllm).
|
| 37 |
+
2. Follow the [RKLLM documentation](https://github.com/airockchip/rknn-llm) to deploy the model.
|
| 38 |
+
3. Use the `rkllm-run` CLI or SDK to perform inference.
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
### License
|
| 42 |
+
This model is released under the **Apache-2.0** license.
|
| 43 |
+
|
| 44 |
+
---
|
orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:95e1d03f0c28a98e6112e0760d4714232b0e5dabb3e239c4566e452796ef1355
|
| 3 |
+
size 7596591734
|
orpheus_3b_0.1_ft_w8a8_RK3588_GGUF_F16.rkllm
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e9bfd72d835e935f1c55754deda88103fa05c7704cc6d0bba3278542819401a
|
| 3 |
+
size 7596587782
|