File size: 1,182 Bytes
9239500 4676ca6 9239500 2fd8041 9239500 4676ca6 9239500 577b1b1 9239500 1e9d92a 9239500 4676ca6 9239500 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
from flask import Flask, request, jsonify
from transformers import pipeline
import torch
app = Flask(__name__)
# ===========================
# LOAD LOCAL/FREE MODEL (HF TOKEN NAHI CHAHIYE)
# ===========================
# Apna model ID yaha daal sakte ho, ya HuggingFace free model
model_id = "google/gemma-2b" # Example free model
print("🔄 Loading model...")
# CPU/GPU device set
device = 0 if torch.cuda.is_available() else -1
# pipeline me device=-1 -> CPU, device=0 -> GPU
ai = pipeline("text-generation", model=model_id, max_new_tokens=200, device=device)
print("✅ Model loaded!")
# ===========================
# CHAT API
# ===========================
@app.route('/chat', methods=['POST'])
def chat():
try:
data = request.get_json()
msg = data.get("message", "")
if not msg:
return jsonify({"error": "No message sent"}), 400
output = ai(msg)[0]["generated_text"]
return jsonify({"reply": output})
except Exception as e:
return jsonify({"error": str(e)}), 500
# ===========================
# RUN SERVER
# ===========================
if __name__ == "__main__":
app.run(host='0.0.0.0', port=7860) |