Cudd1es commited on
Commit
7405997
·
verified ·
1 Parent(s): 445d9bd

switched to QWEN

Browse files

my openai api key was spent 1m tokens by someone yesterday
it made me broken

Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -10,14 +10,14 @@ from sentence_transformers import CrossEncoder
10
  QWEN_MODEL="qwen/qwen3-235b-a22b:free"
11
  DEEPSEEK_MODEL="deepseek/deepseek-chat-v3.1:free"
12
  GPT_OSS_MODEL="openai/gpt-oss-20b:free"
13
- #api_key = os.getenv("OPENROUTER_API_KEY")
14
  #deepseek_key = os.getenv("DEEPSEEK_API_KEY")
15
- #client = OpenAI(base_url="https://openrouter.ai/api/v1", api_key=api_key)
16
 
17
  # open ai method
18
- api_key = os.getenv("OPENAI_API_KEY")
19
- client = OpenAI(api_key=api_key)
20
- GPT_MODEL="gpt-4o"
21
 
22
 
23
 
@@ -25,7 +25,7 @@ collection = load_collection()
25
  encoder = load_encoder()
26
  reranker = CrossEncoder("BAAI/bge-reranker-large")
27
 
28
- def reformulate_query(user_question, model_name=GPT_MODEL):
29
  prompt = f"""你是一个BangDream知识检索助手。请把用户的问题扩写或转写为适合知识库语义检索的检索语句,涵盖所有可能的提问方式或同义关键词。
30
  用户问题:{user_question}
31
  """
@@ -102,7 +102,7 @@ def respond(
102
 
103
  response = ""
104
  stream = client.chat.completions.create(
105
- model=GPT_MODEL,
106
  messages=messages,
107
  temperature=temperature,
108
  max_tokens=max_tokens,
 
10
  QWEN_MODEL="qwen/qwen3-235b-a22b:free"
11
  DEEPSEEK_MODEL="deepseek/deepseek-chat-v3.1:free"
12
  GPT_OSS_MODEL="openai/gpt-oss-20b:free"
13
+ api_key = os.getenv("OPENROUTER_API_KEY")
14
  #deepseek_key = os.getenv("DEEPSEEK_API_KEY")
15
+ client = OpenAI(base_url="https://openrouter.ai/api/v1", api_key=api_key)
16
 
17
  # open ai method
18
+ #api_key = os.getenv("OPENAI_API_KEY")
19
+ #client = OpenAI(api_key=api_key)
20
+ #GPT_MODEL="gpt-4o"
21
 
22
 
23
 
 
25
  encoder = load_encoder()
26
  reranker = CrossEncoder("BAAI/bge-reranker-large")
27
 
28
+ def reformulate_query(user_question, model_name=QWEN_MODEL):
29
  prompt = f"""你是一个BangDream知识检索助手。请把用户的问题扩写或转写为适合知识库语义检索的检索语句,涵盖所有可能的提问方式或同义关键词。
30
  用户问题:{user_question}
31
  """
 
102
 
103
  response = ""
104
  stream = client.chat.completions.create(
105
+ model=QWEN_MODEL,
106
  messages=messages,
107
  temperature=temperature,
108
  max_tokens=max_tokens,