Commit 83ae046
Changed files (4)
src/llm/contexts.py
@@ -67,7 +67,7 @@ async def single_gpt_context(client: Client, message: Message) -> dict:
],
}
"""
- info = parse_msg(message, silent=True)
+ info = parse_msg(message, silent=True, use_cache=False)
role = "assistant" if BOT_TIPS in info["text"] else "user"
if info["mtype"] not in ["text", "photo", "audio", "voice", "video", "document"]:
@@ -79,7 +79,7 @@ async def single_gpt_context(client: Client, message: Message) -> dict:
messages = await client.get_media_group(message.chat.id, message.id) if message.media_group_id else [message]
contexts = []
for msg in messages:
- info = parse_msg(msg, silent=True)
+ info = parse_msg(msg, silent=True, use_cache=False)
sender = info["fwd_full_name"] or info["full_name"]
try:
if info["mtype"] == "photo":
@@ -128,7 +128,7 @@ async def single_gemini_context(client: Client, message: Message, app: genai.Cli
],
}
"""
- info = parse_msg(message, silent=True)
+ info = parse_msg(message, silent=True, use_cache=False)
role = "model" if BOT_TIPS in info["text"] else "user"
if info["mtype"] not in ["text", "photo", "audio", "voice", "video", "document"]:
return {}
@@ -142,7 +142,7 @@ async def single_gemini_context(client: Client, message: Message, app: genai.Cli
messages = await client.get_media_group(message.chat.id, message.id) if message.media_group_id else [message]
parts = []
for msg in messages:
- info = parse_msg(msg, silent=True)
+ info = parse_msg(msg, silent=True, use_cache=False)
sender = info["fwd_full_name"] or info["full_name"]
try:
if info["mtype"] in ["video", "photo", "audio", "voice"] or info["mime_type"] in gemini_mime_types or any(info["file_name"].endswith(ext) for ext in gemini_extensions):
src/llm/gemini.py
@@ -72,7 +72,7 @@ async def gemini_response(
system_prompt (str | None, optional): System prompt. Defaults to None.
silent (bool, optional): Whether to disable progressing. Defaults to False.
"""
- info = parse_msg(message)
+ info = parse_msg(message, silent=True, use_cache=False)
model = GEMINI.TEXT_MODEL if modality == "text" else GEMINI.IMG_MODEL
model_name = GEMINI.TEXT_MODEL_NAME if modality == "text" else GEMINI.IMG_MODEL_NAME
if not GEMINI.API_KEY:
src/llm/gpt.py
@@ -129,7 +129,7 @@ async def gpt_response(
dict: {"texts": str, "thoughts": str, "prefix": str, "model_name": str, "sent_messages": list[Message]}
"""
# ruff: noqa: RET502, RET503
- info = parse_msg(message, use_cache=False)
+ info = parse_msg(message, silent=True, use_cache=False)
# send docs if message == "/ai"
if info["mtype"] == "text" and equal_prefix(info["text"], prefix=PREFIX.GPT):
if not message.reply_to_message: # without reply
src/llm/summary.py
@@ -105,7 +105,7 @@ async def ai_summary(client: Client, message: Message, summary_prefix: str | Non
await send2tg(client, message, texts=HELP, **kwargs)
return
- info = parse_msg(message)
+ info = parse_msg(message, silent=True)
need_summay = startswith_prefix(info["text"], prefix=[PREFIX.AI_SUMMARY])
# replace /combine with /summary, because we need to use `/summary` to match different patterns
info["text"] = re.sub(r"^" + PREFIX.COMBINATION, PREFIX.AI_SUMMARY, info["text"], flags=re.IGNORECASE)