|
3 | 3 | 使用配置管理和更好的错误处理 |
4 | 4 | """ |
5 | 5 | import requests |
6 | | -from typing import Optional |
7 | 6 |
|
8 | 7 | from langchain.agents import tool |
9 | 8 | from langchain_community.utilities import SerpAPIWrapper |
10 | 9 | from langchain_core.output_parsers import JsonOutputParser, StrOutputParser |
11 | 10 | from langchain_core.prompts import ChatPromptTemplate, PromptTemplate |
12 | 11 | from langchain_core.runnables import RunnableLambda |
13 | 12 | from langchain_ollama import OllamaEmbeddings, ChatOllama, OllamaLLM |
| 13 | +from langchain_openai import OpenAI, ChatOpenAI |
14 | 14 | from langchain_qdrant import QdrantVectorStore |
15 | 15 | from qdrant_client import QdrantClient |
16 | 16 |
|
@@ -96,8 +96,12 @@ def bazi_cesuan(query: str) -> str: |
96 | 96 |
|
97 | 97 | # 创建模型 |
98 | 98 | model_config = config.get_model_config() |
99 | | - model = ChatOllama(**model_config, format="json") |
100 | | - |
| 99 | + # model = ChatOllama(**model_config, format="json") |
| 100 | + model = ChatOpenAI( |
| 101 | + model="gpt-4o-mini", |
| 102 | + temperature=0.2, |
| 103 | + ) |
| 104 | + |
101 | 105 | # 构建处理链 |
102 | 106 | chain = prompt | model | parser |
103 | 107 | data = chain.invoke({"query": query}) |
@@ -151,8 +155,12 @@ def jiemeng(query: str) -> str: |
151 | 155 |
|
152 | 156 | # 创建关键词提取模型 |
153 | 157 | model_config = config.get_model_config() |
154 | | - llm = OllamaLLM(**model_config) |
155 | | - |
| 158 | + # llm = OllamaLLM(**model_config) |
| 159 | + llm = OpenAI( |
| 160 | + model="gpt-4o-mini", |
| 161 | + temperature=0.2 |
| 162 | + ) |
| 163 | + |
156 | 164 | # 直接使用统一管理的模板 |
157 | 165 | dream_prompt_template = SystemPrompts.DREAM_KEYWORD_EXTRACTION_PROMPT |
158 | 166 |
|
|
0 commit comments