You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

TunedAI Causal Reasoning Model

Fine-tuned by TunedAI Labs.

What it does

Performs structured causal analysis on business and productivity data. Given a causal question, it reasons through observation, mechanism, projection, and simulation.

Training

  • Base model: Qwen/Qwen2.5-14B-Instruct
  • Fine-tuned by: TunedAI Labs (tunedailabs.com)

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel

base = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-14B-Instruct")
model = PeftModel.from_pretrained(base, "tunedailabs/knapsack-causal-14b")
tokenizer = AutoTokenizer.from_pretrained("tunedailabs/knapsack-causal-14b")

messages = [
    {"role": "system", "content": "You are an expert analyst. When asked causal questions, work through all levels of analysis: patterns in the data, underlying mechanisms, anticipated effects, and counterfactual scenarios."},
    {"role": "user", "content": "Why has this person's email response time increased 40% over the last month?"}
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=1000)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

License

Apache 2.0. Fine-tuned weights by TunedAI Labs. Base model by Alibaba Cloud.

Contact

TunedAI Labs — mark@tunedailabs.com

Downloads last month
36
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tunedailabs/knapsack-causal-7b-merged

Base model

Qwen/Qwen2.5-14B
Quantized
(134)
this model