GLM-4.7-Flash Checkpoint 2500

Best checkpoint from training run - Eval Loss: 0.1504

Model Description

This is checkpoint 2500 from a QLoRA fine-tuning run of GLM-4.7-Flash on a curated dataset of:

  • Agent/tool-use workflows (93.3%)
  • Opus reasoning traces (5.2%)
  • Qwen reasoning data (1.4%)

Training Details

  • Base Model: unsloth/GLM-4.7-Flash (30.4B parameters MoE)
  • Method: QLoRA (4-bit base, LoRA rank r=16)
  • Trainable Params: 1.39%
  • Max Seq Length: 8192
  • Best Eval Loss: 0.1504 (at step 2500)
  • Training Steps: 2500/4998 (50% complete)

Dataset Attribution

This model was fine-tuned on datasets licensed under Apache 2.0:

Primary Sources:

  • Opus-4.6-Reasoning by nohurry | Apache 2.0
  • Qwen3.5-reasoning by Jackrong | Apache 2.0

Usage

This checkpoint contains LoRA adapters. To use:

from unsloth import FastLanguageModel
from peft import PeftModel

# Load base model
model, tokenizer = FastLanguageModel.from_pretrained(
    "unsloth/GLM-4.7-Flash",
    load_in_4bit=True,
)

# Load LoRA adapters
model = PeftModel.from_pretrained(model, "austindixson/glm-4.7-flash-checkpoint-2500")

License

Apache 2.0

Dataset Sources:

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for austindixson/glm-4.7-flash-checkpoint-2500

Finetuned
(18)
this model