Kadoblam commited on
Commit
5789637
·
1 Parent(s): 9572519

Atualiza app.py para usar modelo Qwen-4B-Instruct

Browse files
Files changed (1) hide show
  1. app.py +35 -10
app.py CHANGED
@@ -1,19 +1,44 @@
1
  import os
 
 
2
  from openai import OpenAI
 
 
 
3
 
4
  client = OpenAI(
5
  base_url="https://router.huggingface.co/v1",
6
  api_key=os.environ["HF_TOKEN"],
7
  )
8
 
9
- completion = client.chat.completions.create(
10
- model="meta-llama/Llama-3.1-8B-Instruct:fireworks-ai",
11
- messages=[
12
- {
13
- "role": "user",
14
- "content": "What is the capital of France?"
15
- }
16
- ],
17
- )
 
 
 
 
 
 
 
 
 
 
 
18
 
19
- print(completion.choices[0].message)
 
 
 
 
 
 
 
 
 
 
1
  import os
2
+ from fastapi import FastAPI
3
+ from pydantic import BaseModel
4
  from openai import OpenAI
5
+ import langdetect # pip install langdetect
6
+
7
+ app = FastAPI()
8
 
9
  client = OpenAI(
10
  base_url="https://router.huggingface.co/v1",
11
  api_key=os.environ["HF_TOKEN"],
12
  )
13
 
14
+ class Query(BaseModel):
15
+ question: str
16
+
17
+ @app.get("/")
18
+ async def root():
19
+ return {"message": "API está rodando!"}
20
+
21
+ @app.post("/ask")
22
+ async def ask_model(query: Query):
23
+ try:
24
+ idioma = langdetect.detect(query.question)
25
+ except:
26
+ idioma = "pt"
27
+
28
+ if idioma == "pt":
29
+ system_prompt = "Responda sempre em português, de forma curta e direta, sem explicações longas."
30
+ elif idioma == "en":
31
+ system_prompt = "Answer always in English, briefly and directly, no long explanations."
32
+ else:
33
+ system_prompt = "Responda no mesmo idioma da pergunta, de forma direta e curta."
34
 
35
+ completion = client.chat.completions.create(
36
+ model="meta-llama/Llama-3.1-8B-Instruct:fireworks-ai",
37
+ messages=[
38
+ {"role": "system", "content": system_prompt},
39
+ {"role": "user", "content": query.question}
40
+ ],
41
+ max_tokens=150
42
+ )
43
+ answer = completion.choices[0].message.content
44
+ return {"answer": answer}