SkeptiSTEM-4B-v2 (stageR1) - Merged 16-bit

This is the merged 16-bit checkpoint for SkeptiSTEM-4B-v2 after Stage R1 STEM SFT.

See HallD/SkeptiSTEM-4B-v2-stageR1-lora for the LoRA adapter and training details.

Usage

from unsloth import FastLanguageModel

model, tokenizer = FastLanguageModel.from_pretrained(
    model_name="HallD/SkeptiSTEM-4B-v2-stageR1-merged-16bit",
    max_seq_length=4096,
    load_in_4bit=True,
)
FastLanguageModel.for_inference(model)
Downloads last month
4
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for HallD/SkeptiSTEM-4B-v2-stageR1-merged-16bit

Base model

Qwen/Qwen3-4B-Base
Finetuned
(202)
this model
Adapters
3 models