Commit 83ae046

benny-dou <60535774+benny-dou@users.noreply.github.com>
2025-06-17 15:23:12
chore(parser): ignore cache in some cases
1 parent b346a67
Changed files (4)
src/llm/contexts.py
@@ -67,7 +67,7 @@ async def single_gpt_context(client: Client, message: Message) -> dict:
         ],
     }
     """
-    info = parse_msg(message, silent=True)
+    info = parse_msg(message, silent=True, use_cache=False)
     role = "assistant" if BOT_TIPS in info["text"] else "user"
 
     if info["mtype"] not in ["text", "photo", "audio", "voice", "video", "document"]:
@@ -79,7 +79,7 @@ async def single_gpt_context(client: Client, message: Message) -> dict:
     messages = await client.get_media_group(message.chat.id, message.id) if message.media_group_id else [message]
     contexts = []
     for msg in messages:
-        info = parse_msg(msg, silent=True)
+        info = parse_msg(msg, silent=True, use_cache=False)
         sender = info["fwd_full_name"] or info["full_name"]
         try:
             if info["mtype"] == "photo":
@@ -128,7 +128,7 @@ async def single_gemini_context(client: Client, message: Message, app: genai.Cli
         ],
     }
     """
-    info = parse_msg(message, silent=True)
+    info = parse_msg(message, silent=True, use_cache=False)
     role = "model" if BOT_TIPS in info["text"] else "user"
     if info["mtype"] not in ["text", "photo", "audio", "voice", "video", "document"]:
         return {}
@@ -142,7 +142,7 @@ async def single_gemini_context(client: Client, message: Message, app: genai.Cli
     messages = await client.get_media_group(message.chat.id, message.id) if message.media_group_id else [message]
     parts = []
     for msg in messages:
-        info = parse_msg(msg, silent=True)
+        info = parse_msg(msg, silent=True, use_cache=False)
         sender = info["fwd_full_name"] or info["full_name"]
         try:
             if info["mtype"] in ["video", "photo", "audio", "voice"] or info["mime_type"] in gemini_mime_types or any(info["file_name"].endswith(ext) for ext in gemini_extensions):
src/llm/gemini.py
@@ -72,7 +72,7 @@ async def gemini_response(
         system_prompt (str | None, optional): System prompt. Defaults to None.
         silent (bool, optional): Whether to disable progressing. Defaults to False.
     """
-    info = parse_msg(message)
+    info = parse_msg(message, silent=True, use_cache=False)
     model = GEMINI.TEXT_MODEL if modality == "text" else GEMINI.IMG_MODEL
     model_name = GEMINI.TEXT_MODEL_NAME if modality == "text" else GEMINI.IMG_MODEL_NAME
     if not GEMINI.API_KEY:
src/llm/gpt.py
@@ -129,7 +129,7 @@ async def gpt_response(
         dict: {"texts": str, "thoughts": str, "prefix": str, "model_name": str, "sent_messages": list[Message]}
     """
     # ruff: noqa: RET502, RET503
-    info = parse_msg(message, use_cache=False)
+    info = parse_msg(message, silent=True, use_cache=False)
     # send docs if message == "/ai"
     if info["mtype"] == "text" and equal_prefix(info["text"], prefix=PREFIX.GPT):
         if not message.reply_to_message:  # without reply
src/llm/summary.py
@@ -105,7 +105,7 @@ async def ai_summary(client: Client, message: Message, summary_prefix: str | Non
         await send2tg(client, message, texts=HELP, **kwargs)
         return
 
-    info = parse_msg(message)
+    info = parse_msg(message, silent=True)
     need_summay = startswith_prefix(info["text"], prefix=[PREFIX.AI_SUMMARY])
     # replace /combine with /summary, because we need to use `/summary` to match different patterns
     info["text"] = re.sub(r"^" + PREFIX.COMBINATION, PREFIX.AI_SUMMARY, info["text"], flags=re.IGNORECASE)