Math & Reasoning LLMs
Collection
Math & Reasoning LLMs by Harsha901
•
6 items
•
Updated
This Qwen2.5-7B model was fine-tuned using Unsloth for faster and more memory-efficient training, together with Hugging Face’s TRL library for supervised fine-tuning.
This is an instruction-tuned causal language model based on Qwen2.5-7B, designed to follow user prompts accurately and generate coherent, high-quality responses.
The model preserves the general-purpose strengths of Qwen2.5 while benefiting from domain-focused supervised fine-tuning.
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "Harsha901/<YOUR-MODEL-NAME>"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto",
torch_dtype="auto"
)
Released under the Apache 2.0 License, consistent with the base Qwen2.5 model.