|
1 | 1 | try:
|
2 |
| - from ..llm_factory import OpenAILLMs |
| 2 | + from ..llm_factory import OpenAILLMs, GoogleAILLMs |
3 | 3 | from .informational_prompts import \
|
4 | 4 | informational_role_prompt, conv_pref_prompt, update_conv_pref_prompt, summary_prompt, update_summary_prompt, summary_system_prompt
|
5 | 5 | from ..utils.types import InvokeAgentResponseType
|
6 | 6 | except ImportError:
|
7 |
| - from src.agents.llm_factory import OpenAILLMs |
| 7 | + from src.agents.llm_factory import OpenAILLMs, GoogleAILLMs |
8 | 8 | from src.agents.informational_agent.informational_prompts import \
|
9 | 9 | informational_role_prompt, conv_pref_prompt, update_conv_pref_prompt, summary_prompt, update_summary_prompt, summary_system_prompt
|
10 | 10 | from src.agents.utils.types import InvokeAgentResponseType
|
@@ -37,7 +37,7 @@ class State(TypedDict):
|
37 | 37 |
|
38 | 38 | class InformationalAgent:
|
39 | 39 | def __init__(self):
|
40 |
| - llm = OpenAILLMs(temperature=0.25) |
| 40 | + llm = GoogleAILLMs() |
41 | 41 | self.llm = llm.get_llm()
|
42 | 42 | summarisation_llm = OpenAILLMs()
|
43 | 43 | self.summarisation_llm = summarisation_llm.get_llm()
|
@@ -135,7 +135,6 @@ def summarize_conversation(self, state: State, config: RunnableConfig) -> dict:
|
135 | 135 | delete_messages: list[AllMessageTypes] = [RemoveMessage(id=m.id) for m in state["messages"][:-3]]
|
136 | 136 |
|
137 | 137 | return {"summary": summary_response.content, "conversationalStyle": conversationalStyle_response.content, "messages": delete_messages}
|
138 |
| - # return {"summary": summary_response.content, "messages": delete_messages} |
139 | 138 |
|
140 | 139 | def should_summarize(self, state: State) -> str:
|
141 | 140 | """
|
|
0 commit comments