Harsha901 commited on
Commit
35354db
·
verified ·
1 Parent(s): 051ad97

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -9
README.md CHANGED
@@ -1,24 +1,80 @@
1
  ---
2
- base_model:
3
- - unsloth/Qwen2.5-7B-Instruct
4
- - Qwen/Qwen2.5-7B-Instruct
5
  tags:
6
  - text-generation-inference
7
  - transformers
8
  - unsloth
9
  - qwen2
10
  - trl
 
 
 
11
  license: apache-2.0
12
  language:
13
  - en
 
14
  ---
15
 
16
- # Uploaded model
17
 
18
- - **Developed by:** Harsha901
19
- - **License:** apache-2.0
20
- - **Finetuned from model :** unsloth/Qwen2.5-7B-Instruct
21
 
22
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
23
 
24
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: unsloth/Qwen2.5-7B-Instruct
 
 
3
  tags:
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - qwen2
8
  - trl
9
+ - instruction-tuned
10
+ - supervised-finetuning
11
+ - causal-lm
12
  license: apache-2.0
13
  language:
14
  - en
15
+ pipeline_tag: text-generation
16
  ---
17
 
18
+ # Uploaded Model
19
 
20
+ - **Developed by:** Harsha901
21
+ - **License:** Apache-2.0
22
+ - **Finetuned from model:** unsloth/Qwen2.5-7B-Instruct
23
 
24
+ This Qwen2.5-7B model was fine-tuned using **Unsloth** for faster and more memory-efficient training, together with Hugging Face’s **TRL** library for supervised fine-tuning.
25
 
26
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
27
+
28
+ ---
29
+
30
+ ## Model Overview
31
+
32
+ This is an **instruction-tuned causal language model** based on **Qwen2.5-7B**, designed to follow user prompts accurately and generate coherent, high-quality responses.
33
+
34
+ The model preserves the general-purpose strengths of Qwen2.5 while benefiting from domain-focused supervised fine-tuning.
35
+
36
+ ---
37
+
38
+ ## Training Details
39
+
40
+ - **Base model:** Qwen2.5-7B-Instruct (Unsloth variant)
41
+ - **Fine-tuning method:** Supervised Fine-Tuning (SFT)
42
+ - **Frameworks:** Hugging Face Transformers + TRL
43
+ - **Acceleration:** Unsloth (2× faster training, reduced VRAM usage)
44
+ - **Precision:** FP16 / BF16 (hardware dependent)
45
+
46
+ ---
47
+ ## Usage
48
+
49
+ ```python
50
+ from transformers import AutoModelForCausalLM, AutoTokenizer
51
+
52
+ model_id = "Harsha901/<YOUR-MODEL-NAME>"
53
+
54
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
55
+ model = AutoModelForCausalLM.from_pretrained(
56
+ model_id,
57
+ device_map="auto",
58
+ torch_dtype="auto"
59
+ )
60
+ ````
61
+ ---
62
+
63
+ ## Limitations
64
+
65
+ * Outputs may contain factual or reasoning errors
66
+ * Not intended for high-stakes or safety-critical applications
67
+ * Performance depends on prompt quality and context length
68
+
69
+ ---
70
+ ## License
71
+
72
+ Released under the **Apache 2.0 License**, consistent with the base Qwen2.5 model.
73
+
74
+ ---
75
+ ## Acknowledgements
76
+
77
+ * **Qwen Team** for the Qwen2.5 base model
78
+ * **Unsloth** for efficient fine-tuning optimizations
79
+ * **Hugging Face** for the training and hosting ecosystem
80
+ ```