SkeptiSTEM-4B-v2 (stageR1) - Merged 16-bit
This is the merged 16-bit checkpoint for SkeptiSTEM-4B-v2 after Stage R1 STEM SFT.
See HallD/SkeptiSTEM-4B-v2-stageR1-lora for the LoRA adapter and training details.
Usage
from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
model_name="HallD/SkeptiSTEM-4B-v2-stageR1-merged-16bit",
max_seq_length=4096,
load_in_4bit=True,
)
FastLanguageModel.for_inference(model)
- Downloads last month
- 4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support