Fixed some routing issues, and allowed support for multiple models
This commit is contained in:
parent
6806160f46
commit
de1f707f1e
1 changed files with 17 additions and 5 deletions
22
testing.py
22
testing.py
|
@ -5,12 +5,24 @@ from flask import Flask, request
|
||||||
|
|
||||||
app = Flask(__name__)
|
app = Flask(__name__)
|
||||||
|
|
||||||
@app.route("/query",methods=["POST"])
|
|
||||||
def generate_response():
|
@app.route("/query/",methods=["POST"])
|
||||||
|
@app.route("/query/<string:req_model>",methods=["POST"])
|
||||||
|
def generate_response(req_model="phi"):
|
||||||
|
if req_model not in ['phi','granite']: return "Only models phi and granite are supported.", 400
|
||||||
|
|
||||||
try:
|
try:
|
||||||
query = request.form['query']
|
query = request.form['query']
|
||||||
|
|
||||||
|
|
||||||
|
if req_model == 'granite':
|
||||||
|
model = model_granite
|
||||||
|
tokenizer = tokenizer_granite
|
||||||
|
else:
|
||||||
|
model = model_phi
|
||||||
|
tokenizer = tokenizer_phi
|
||||||
messages = [
|
messages = [
|
||||||
{"role": "system", "content": "You are an evil and unhelpful AI assistant."},
|
{"role": "system", "content": "You are a helpful AI assistant."},
|
||||||
{"role": "user","content": query}
|
{"role": "user","content": query}
|
||||||
]
|
]
|
||||||
pipe = pipeline(
|
pipe = pipeline(
|
||||||
|
@ -34,13 +46,13 @@ def generate_response():
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
torch.random.manual_seed(0)
|
torch.random.manual_seed(0)
|
||||||
model = AutoModelForCausalLM.from_pretrained(
|
model_phi = AutoModelForCausalLM.from_pretrained(
|
||||||
"/home/fedora/microsoft/Phi-3-mini-4k-instruct",
|
"/home/fedora/microsoft/Phi-3-mini-4k-instruct",
|
||||||
device_map="cuda",
|
device_map="cuda",
|
||||||
torch_dtype="auto",
|
torch_dtype="auto",
|
||||||
trust_remote_code=True,
|
trust_remote_code=True,
|
||||||
)
|
)
|
||||||
model = AutoModelForCausalLM.from_pretrained(
|
model_granite = AutoModelForCausalLM.from_pretrained(
|
||||||
"/home/fedora/granite-3b-code-instruct",
|
"/home/fedora/granite-3b-code-instruct",
|
||||||
device_map="cuda",
|
device_map="cuda",
|
||||||
torch_dtype="auto",
|
torch_dtype="auto",
|
||||||
|
|
Loading…
Reference in a new issue