qwen35-27b-v2-stage1
This model is a fine-tuned version of llmfan46/Qwen3.5-27B-heretic-v2.
Content transfer stage — lower-quality but content-valuable data trained with lower LR and high rank / low alpha ratio. Plain text (all tokens trained), single epoch, 2048 context.
W&B run: https://wandb.ai/cooawoo-personal/huggingface/runs/4cesvope
Dataset
~28.4M tokens total from 8 datasets, split from 9,054 samples into 20,393 chunks at 2048 max length.
Training Configs
train.yaml
# Qwen3.5-27B Step 1 — Content transfer (lower-quality data)
#
# Lower-quality but content-valuable data. Content transfer via
# lower LR + high rank / low alpha ratio.
#
# ~45M tokens, single epoch, 2048 context (rosier 1/4 sample)
# Base model: llmfan46/Qwen3.5-27B-heretic-v2
accelerate_config: /home/aibox/loft-t5/configs/qwen35-27b-step1/accelerate_config.yaml
model_name_or_path: /home/aibox/.cache/huggingface/hub/models--llmfan46--Qwen3.5-27B-heretic-v2/snapshots/dbb3412746fa6b27523c5cd74c534fedf0e3355d
data_config: /home/aibox/loft-t5/configs/qwen35-27b-step1/data.yaml
prepared_dataset: /home/aibox/loft-t5/runs/qwen35-27b-step1/prepared
output_dir: /home/aibox/loft-t5/runs/qwen35-27b-step1
# Precision & memory
bf16: true
gradient_checkpointing: true
use_cce: true
# Context & batching
max_length: 2048
per_device_train_batch_size: 1
gradient_accumulation_steps: 4
# QLoRA — larger rank with low alpha for content transfer
use_peft: true
load_in_4bit: true
model_parallel: true
lora_r: 64
lora_alpha: 8
lora_dropout: 0.0
use_rslora: true
# Optimizer & schedule — lower LR for content without style
learning_rate: 5.0e-5
lr_scheduler_type: cosine
warmup_ratio: 0.03
weight_decay: 0.01
max_grad_norm: 1.0
num_train_epochs: 1
# Logging & saving
logging_steps: 1
disable_tqdm: true
save_strategy: steps
save_steps: 1000
save_total_limit: 5
report_to: wandb
run_name: qwen35-27b-step1
data.yaml
# Data config — Qwen3.5-27B Step 1 (content transfer, lower-quality data)
# ~24M tokens total, single epoch, 2048 context
datasets:
# Rosier inf strict — 1/8 random sample (~10M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/rosier_inf_strict_halved.jsonl
type: text
columns: [text]
truncation_strategy: split
# Erotica quality cleaned — half sample (~6.7M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/erotica_quality_cleaned_halved.jsonl
type: text
columns: [text]
truncation_strategy: split
# Erotic books filtered — longer-form erotica (~2.8M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/erotic_books_filtered.json
type: text
columns: [text]
truncation_strategy: split
# Springdragon — half sample (~3.4M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/springdragon_processed_halved.jsonl
type: text
columns: [text]
truncation_strategy: split
# Floyd — half sample (~3.9M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/floyd_processed_halved.jsonl
type: text
columns: [text]
truncation_strategy: split
# Brainrot chatlog — internet chat style (~1.1M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/brainrot_chatlog.jsonl
type: text
columns: [text]
truncation_strategy: split
# Wrecklora — half sample (~6.3M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/wrecklora_processed_halved.jsonl
type: text
columns: [text]
truncation_strategy: split
# Disco Elysium chat — interactive fiction dialogue (~0.5M tokens)
- path: /home/aibox/loft-t5/configs/qwen35-27b-step1/data/disco_chat.jsonl
type: text
columns: [text]
truncation_strategy: split
# Shuffle settings
shuffle_datasets: true
shuffle_combined: true
shuffle_seed: 42
# No eval split — single epoch
eval_split: 0
split_seed: 42
# Plain text — train on all tokens
assistant_only_loss: false
accelerate_config.yaml
compute_environment: LOCAL_MACHINE
distributed_type: 'NO'
num_processes: 1
mixed_precision: 'no'
Framework versions
- PEFT 0.18.1
- Loft: 0.1.0
- Transformers: 5.2.0
- Pytorch: 2.6.0
- Datasets: 4.6.1
- Tokenizers: 0.22.2
- Downloads last month
- 66