Uploaded Model

  • Developed by: Harsha901
  • License: Apache-2.0
  • Finetuned from model: unsloth/Qwen2.5-7B-Instruct

This Qwen2.5-7B model was fine-tuned using Unsloth for faster and more memory-efficient training, together with Hugging Face’s TRL library for supervised fine-tuning.


Model Overview

This is an instruction-tuned causal language model based on Qwen2.5-7B, designed to follow user prompts accurately and generate coherent, high-quality responses.

The model preserves the general-purpose strengths of Qwen2.5 while benefiting from domain-focused supervised fine-tuning.


Training Details

  • Base model: Qwen2.5-7B-Instruct (Unsloth variant)
  • Fine-tuning method: Supervised Fine-Tuning (SFT)
  • Frameworks: Hugging Face Transformers + TRL
  • Acceleration: Unsloth (2× faster training, reduced VRAM usage)
  • Precision: FP16 / BF16 (hardware dependent)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "Harsha901/<YOUR-MODEL-NAME>"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    device_map="auto",
    torch_dtype="auto"
)

Limitations

  • Outputs may contain factual or reasoning errors
  • Not intended for high-stakes or safety-critical applications
  • Performance depends on prompt quality and context length

License

Released under the Apache 2.0 License, consistent with the base Qwen2.5 model.


Acknowledgements

  • Qwen Team for the Qwen2.5 base model
  • Unsloth for efficient fine-tuning optimizations
  • Hugging Face for the training and hosting ecosystem

Downloads last month
16
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Harsha901/Qwen2.5-7B-Inst-Math-Reasoning-SFT

Base model

Qwen/Qwen2.5-7B
Finetuned
(1294)
this model

Collection including Harsha901/Qwen2.5-7B-Inst-Math-Reasoning-SFT