From fbc50082599eea86e9101803be58f8a8b22bf971 Mon Sep 17 00:00:00 2001 From: pablodanswer Date: Mon, 16 Sep 2024 12:26:05 -0700 Subject: [PATCH] logs --- backend/danswer/chat/process_message.py | 1 + backend/danswer/llm/answering/answer.py | 18 ++++-------------- backend/danswer/tools/tool_runner.py | 7 ++----- web/src/app/chat/ChatPage.tsx | 1 + 4 files changed, 8 insertions(+), 19 deletions(-) diff --git a/backend/danswer/chat/process_message.py b/backend/danswer/chat/process_message.py index c9f40ca95..67bc38bb4 100644 --- a/backend/danswer/chat/process_message.py +++ b/backend/danswer/chat/process_message.py @@ -946,4 +946,5 @@ def stream_chat_message( is_connected=is_connected, ) for obj in objects: + print(obj) yield get_json_line(obj.model_dump()) diff --git a/backend/danswer/llm/answering/answer.py b/backend/danswer/llm/answering/answer.py index 8f9043268..073b83b81 100644 --- a/backend/danswer/llm/answering/answer.py +++ b/backend/danswer/llm/answering/answer.py @@ -272,9 +272,9 @@ class Answer: stop_reason=StreamStopReason.CONTEXT_LENGTH ) - if not tool_call_chunk: - logger.info("Skipped tool call but generated message") - return + if not tool_call_chunk: + logger.info("Skipped tool call but generated message") + return tool_call_requests = tool_call_chunk.tool_calls @@ -328,7 +328,6 @@ class Answer: self._update_prompt_builder_for_search_tool(prompt_builder, []) elif tool.name == ImageGenerationTool._NAME: - print("\n----\nUpdating image prompt user message\n----\n") img_urls = [ img_generation_result["url"] for img_generation_result in tool_runner.tool_final_result().tool_result @@ -339,8 +338,6 @@ class Answer: ) ) - print("now stremign wie fianl results") - yield tool_runner.tool_final_result() # Update message history with tool call and response @@ -379,17 +376,11 @@ class Answer: ) ) - print("\n----\nBuilding final prompt with Tool call summary\n----\n") - # Generate response based on updated message history prompt = prompt_builder.build(tool_call_summary=tool_call_summary) response_content = "" - for content in self._process_llm_stream( - prompt=prompt, - tools=None - # tools=[tool.tool_definition() for tool in self.tools], - ): + for content in self._process_llm_stream(prompt=prompt, tools=None): if isinstance(content, str): response_content += content yield content @@ -765,5 +756,4 @@ class Answer: if not self.is_connected(): logger.debug("Answer stream has been cancelled") self._is_cancelled = not self.is_connected() - return self._is_cancelled diff --git a/backend/danswer/tools/tool_runner.py b/backend/danswer/tools/tool_runner.py index 7e0df959d..17b1de799 100644 --- a/backend/danswer/tools/tool_runner.py +++ b/backend/danswer/tools/tool_runner.py @@ -12,7 +12,7 @@ from danswer.utils.threadpool_concurrency import run_functions_tuples_in_paralle class ToolRunner: - def __init__(self, tool: Tool, args: dict[str, Any], llm: LLM): + def __init__(self, tool: Tool, args: dict[str, Any], llm: LLM | None = None): self.tool = tool self.args = args self._llm = llm @@ -49,12 +49,9 @@ class ToolRunner: def check_which_tools_should_run_for_non_tool_calling_llm( tools: list[Tool], query: str, history: list[PreviousMessage], llm: LLM ) -> list[dict[str, Any] | None]: - print(len(tools)) tool_args_list: list[tuple[Callable[..., Any], tuple[Any, ...]]] = [ (tool.get_args_for_non_tool_calling_llm, (query, history, llm)) for tool in tools ] - print(tool_args_list) - print(tools) - print(len(tool_args_list)) + return run_functions_tuples_in_parallel(tool_args_list) diff --git a/web/src/app/chat/ChatPage.tsx b/web/src/app/chat/ChatPage.tsx index 0129ec3a2..c37543574 100644 --- a/web/src/app/chat/ChatPage.tsx +++ b/web/src/app/chat/ChatPage.tsx @@ -1142,6 +1142,7 @@ export function ChatPage({ if (!packet) { continue; } + console.log(packet); if (!initialFetchDetails) { if (!Object.hasOwn(packet, "user_message_id")) {