parallel-ttt-cot-qwen3-4b-v1
LoRA adapter checkpoint exported from epoch=0-step=4200.ckpt on top of Qwen/Qwen3-4B-Instruct-2507.
This repo intentionally stores the adapter separately from the base model so it can be resumed for later LoRA finetuning.
Load for inference or further finetuning
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
base_model_name = "Qwen/Qwen3-4B-Instruct-2507"
adapter_name = "zechen-nlp/parallel-ttt-cot-qwen3-4b-v1"
tokenizer = AutoTokenizer.from_pretrained(adapter_name)
base_model = AutoModelForCausalLM.from_pretrained(
base_model_name,
torch_dtype="auto",
trust_remote_code=True,
)
model = PeftModel.from_pretrained(base_model, adapter_name)
Included files
adapter_config.jsonadapter_model.safetensors- tokenizer files copied from the base Qwen3 repo, including
tokenizer_config.jsonwith the chat template generation_config.json
- Downloads last month
- 15
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for zechen-nlp/parallel-ttt-cot-qwen3-4b-v1
Base model
Qwen/Qwen3-4B-Instruct-2507