mirror of
https://github.com/LiteyukiStudio/nonebot-plugin-marshoai.git
synced 2024-11-26 23:05:04 +08:00
不同会话间的上下文分离,优化帮助文本
This commit is contained in:
parent
6601b236b7
commit
e408a48e42
20
__init__.py
20
__init__.py
@ -1,20 +1,20 @@
|
||||
from nonebot.plugin import PluginMetadata, inherit_supported_adapters, require
|
||||
require("nonebot_plugin_htmlrender")
|
||||
require("nonebot_plugin_alconna")
|
||||
from .azure import *
|
||||
from nonebot import get_driver
|
||||
#from .config import ConfigModel
|
||||
usage = """MarshoAI Alpha by Asankilp
|
||||
用法:
|
||||
marsho <聊天内容>
|
||||
与 Marsho 进行对话。当模型为gpt时,可以带上图片进行对话。
|
||||
changemodel <模型名>
|
||||
切换 AI 模型。仅超级用户可用。
|
||||
reset
|
||||
重置上下文。仅超级用户可用。
|
||||
marsho <聊天内容> : 与 Marsho 进行对话。当模型为 GPT-4o(-mini) 等时,可以带上图片进行对话。
|
||||
reset : 重置当前会话的上下文。
|
||||
超级用户命令:
|
||||
changemodel <模型名> : 切换全局 AI 模型。
|
||||
contexts : 返回当前会话的上下文列表。 ※当上下文包含图片时,不要使用此命令。
|
||||
praises : 返回夸赞名单的提示词。
|
||||
usermsg <消息> : 往当前会话添加用户消息(UserMessage)。
|
||||
assistantmsg <消息> : 往当前会话添加助手消息(AssistantMessage)。
|
||||
注意事项:
|
||||
当 Marsho 回复消息为None或以content_filter开头的错误信息时,表示该消息被内容过滤器过滤,请调整你的聊天内容确保其合规。
|
||||
当回复以RateLimitReached开头的错误信息时,该 AI 模型的次数配额已用尽,请联系Bot管理员。
|
||||
- 当 Marsho 回复消息为None或以content_filter开头的错误信息时,表示该消息被内容过滤器过滤,请调整你的聊天内容确保其合规。
|
||||
- 当回复以RateLimitReached开头的错误信息时,该 AI 模型的次数配额已用尽,请联系Bot管理员。
|
||||
※本AI的回答"按原样"提供,不提供任何担保。AI也会犯错,请仔细甄别回答的准确性。"""
|
||||
__author__ = "Asankilp"
|
||||
__plugin_meta__ = PluginMetadata(
|
||||
|
48
azure.py
48
azure.py
@ -1,10 +1,9 @@
|
||||
from nonebot.typing import T_State
|
||||
from nonebot import on_command
|
||||
from nonebot.adapters import Message
|
||||
from nonebot.params import CommandArg
|
||||
from nonebot.permission import SUPERUSER
|
||||
#from .acgnapis import *
|
||||
from nonebot_plugin_alconna import on_alconna
|
||||
from nonebot_plugin_alconna import on_alconna, MsgTarget
|
||||
from nonebot_plugin_alconna.uniseg import UniMessage, UniMsg
|
||||
from arclet.alconna import Alconna, Args, AllParam
|
||||
from .util import *
|
||||
@ -16,8 +15,8 @@ from .__init__ import __plugin_meta__
|
||||
from .config import config
|
||||
from .models import MarshoContext
|
||||
changemodel_cmd = on_command("changemodel",permission=SUPERUSER)
|
||||
resetmem_cmd = on_command("reset",permission=SUPERUSER)
|
||||
setprompt_cmd = on_command("prompt",permission=SUPERUSER)
|
||||
resetmem_cmd = on_command("reset")
|
||||
#setprompt_cmd = on_command("prompt",permission=SUPERUSER)
|
||||
praises_cmd = on_command("praises",permission=SUPERUSER)
|
||||
add_usermsg_cmd = on_command("usermsg",permission=SUPERUSER)
|
||||
add_assistantmsg_cmd = on_command("assistantmsg",permission=SUPERUSER)
|
||||
@ -30,20 +29,17 @@ marsho_cmd = on_alconna(
|
||||
)
|
||||
model_name = config.marshoai_default_model
|
||||
context = MarshoContext()
|
||||
context_limit = 50
|
||||
|
||||
|
||||
|
||||
@add_usermsg_cmd.handle()
|
||||
async def add_usermsg(arg: Message = CommandArg()):
|
||||
async def add_usermsg(target: MsgTarget, arg: Message = CommandArg()):
|
||||
if msg := arg.extract_plain_text():
|
||||
context.append(UserMessage(content=msg))
|
||||
context.append(UserMessage(content=msg), target.id, target.private)
|
||||
await UniMessage("已添加用户消息").send()
|
||||
|
||||
@add_assistantmsg_cmd.handle()
|
||||
async def add_assistantmsg(arg: Message = CommandArg()):
|
||||
async def add_assistantmsg(target: MsgTarget, arg: Message = CommandArg()):
|
||||
if msg := arg.extract_plain_text():
|
||||
context.append(AssistantMessage(content=msg))
|
||||
context.append(AssistantMessage(content=msg), target.id, target.private)
|
||||
await UniMessage("已添加助手消息").send()
|
||||
|
||||
@praises_cmd.handle()
|
||||
@ -51,8 +47,8 @@ async def praises():
|
||||
await UniMessage(build_praises()).send()
|
||||
|
||||
@contexts_cmd.handle()
|
||||
async def contexts():
|
||||
await UniMessage(str(context.build()[1:])).send()
|
||||
async def contexts(target: MsgTarget):
|
||||
await UniMessage(str(context.build(target.id, target.private)[1:])).send()
|
||||
|
||||
# @setprompt_cmd.handle() #用不了了
|
||||
# async def setprompt(arg: Message = CommandArg()):
|
||||
@ -67,9 +63,8 @@ async def contexts():
|
||||
|
||||
|
||||
@resetmem_cmd.handle()
|
||||
async def resetmem_cmd():
|
||||
context.reset()
|
||||
context.resetcount()
|
||||
async def resetmem(target: MsgTarget):
|
||||
context.reset(target.id, target.private)
|
||||
await resetmem_cmd.finish("上下文已重置")
|
||||
|
||||
@changemodel_cmd.handle()
|
||||
@ -80,6 +75,7 @@ async def changemodel(arg : Message = CommandArg()):
|
||||
await changemodel_cmd.finish("已切换")
|
||||
@marsho_cmd.handle()
|
||||
async def marsho(
|
||||
target: MsgTarget,
|
||||
message: UniMsg,
|
||||
text = None
|
||||
):
|
||||
@ -94,14 +90,11 @@ async def marsho(
|
||||
await UniMessage(
|
||||
__plugin_meta__.usage+"\n当前使用的模型:"+model_name).send()
|
||||
return
|
||||
if context.count >= context_limit:
|
||||
await UniMessage("上下文数量达到阈值。已自动重置上下文。").send()
|
||||
context.reset()
|
||||
context.resetcount()
|
||||
# await UniMessage(str(text)).send()
|
||||
try:
|
||||
is_support_image_model = model_name.lower() in config.marshoai_support_image_models
|
||||
usermsg = [] if is_support_image_model else ""
|
||||
marsho_string_removed = False
|
||||
for i in message:
|
||||
if i.type == "image":
|
||||
if is_support_image_model:
|
||||
@ -113,20 +106,25 @@ async def marsho(
|
||||
else:
|
||||
await UniMessage("*此模型不支持图片处理。").send()
|
||||
elif i.type == "text":
|
||||
if not marsho_string_removed:
|
||||
# 去掉最前面的"marsho "字符串
|
||||
clean_text = i.data["text"].lstrip("marsho ")
|
||||
marsho_string_removed = True # 标记文本已处理
|
||||
else:
|
||||
clean_text = i.data["text"]
|
||||
if is_support_image_model:
|
||||
usermsg.append(TextContentItem(text=i.data["text"]))
|
||||
usermsg.append(TextContentItem(text=clean_text))
|
||||
else:
|
||||
usermsg += str(i.data["text"])
|
||||
response = await client.complete(
|
||||
messages=context.build()+[UserMessage(content=usermsg)],
|
||||
messages=context.build(target.id, target.private)+[UserMessage(content=usermsg)],
|
||||
model=model_name
|
||||
)
|
||||
#await UniMessage(str(response)).send()
|
||||
choice = response.choices[0]
|
||||
if choice["finish_reason"] == CompletionsFinishReason.STOPPED:
|
||||
context.append(UserMessage(content=usermsg))
|
||||
context.append(choice.message)
|
||||
context.addcount()
|
||||
context.append(UserMessage(content=usermsg), target.id, target.private)
|
||||
context.append(choice.message, target.id, target.private)
|
||||
elif choice["finish_reason"] == CompletionsFinishReason.CONTENT_FILTERED:
|
||||
await UniMessage("*已被内容过滤器过滤。*").send()
|
||||
#await UniMessage(str(choice)).send()
|
||||
|
@ -1,4 +1,3 @@
|
||||
from pathlib import Path
|
||||
from pydantic import BaseModel
|
||||
from nonebot import get_plugin_config
|
||||
|
||||
|
37
models.py
37
models.py
@ -1,35 +1,40 @@
|
||||
from .util import *
|
||||
|
||||
class MarshoContext:
|
||||
"""
|
||||
Marsho 的上下文类
|
||||
"""
|
||||
def __init__(self):
|
||||
self.contents = []
|
||||
self.count = 0
|
||||
self.contents = {
|
||||
"private": {},
|
||||
"non-private": {}
|
||||
}
|
||||
|
||||
def append(self, content):
|
||||
def _get_target_dict(self, is_private):
|
||||
return self.contents["private"] if is_private else self.contents["non-private"]
|
||||
|
||||
def append(self, content, target_id, is_private):
|
||||
"""
|
||||
往上下文中添加消息
|
||||
Args:
|
||||
content: 消息
|
||||
"""
|
||||
self.contents.append(content)
|
||||
target_dict = self._get_target_dict(is_private)
|
||||
if target_id not in target_dict:
|
||||
target_dict[target_id] = []
|
||||
target_dict[target_id].append(content)
|
||||
|
||||
def reset(self):
|
||||
def reset(self, target_id, is_private):
|
||||
"""
|
||||
重置上下文
|
||||
"""
|
||||
self.contents.clear()
|
||||
target_dict = self._get_target_dict(is_private)
|
||||
target_dict[target_id].clear()
|
||||
|
||||
def addcount(self, num = 1):
|
||||
self.count += num
|
||||
|
||||
def resetcount(self):
|
||||
self.count = 0
|
||||
|
||||
def build(self):
|
||||
def build(self, target_id, is_private):
|
||||
"""
|
||||
构建返回的上下文,其中包括系统消息
|
||||
"""
|
||||
spell = get_prompt()
|
||||
return [spell] + self.contents
|
||||
target_dict = self._get_target_dict(is_private)
|
||||
if target_id not in target_dict:
|
||||
target_dict[target_id] = []
|
||||
return [spell] + target_dict[target_id]
|
4
util.py
4
util.py
@ -1,10 +1,8 @@
|
||||
import base64
|
||||
import mimetypes
|
||||
import random
|
||||
import os
|
||||
import json
|
||||
import httpx
|
||||
from pathlib import Path
|
||||
from datetime import datetime
|
||||
from zhDateTime import DateTime
|
||||
from azure.ai.inference.models import SystemMessage
|
||||
@ -34,7 +32,7 @@ def get_praises():
|
||||
if not os.path.exists("praises.json"):
|
||||
init_data = {
|
||||
"like": [
|
||||
{"name":"Asankilp","advantages":"赋予了Marsho猫娘人格,使用vim为Marsho写了许多代码,使Marsho更加可爱"}
|
||||
{"name":"Asankilp","advantages":"赋予了Marsho猫娘人格,使用vim与vscode为Marsho写了许多代码,使Marsho更加可爱"}
|
||||
]
|
||||
}
|
||||
with open(filename,"w",encoding="utf-8") as f:
|
||||
|
Loading…
Reference in New Issue
Block a user