feat: 添加网络健康检查,完善LLM

This commit is contained in:
feie9454 2025-10-17 08:21:25 +08:00
parent 3506ee9bd0
commit 0ace2929a5
10 changed files with 406 additions and 83 deletions

View File

@ -7,6 +7,7 @@
"@types/cors": "^2.8.19",
"@types/multer": "^2.0.0",
"@types/node-persist": "^3.1.8",
"chalk": "^5",
"cors": "^2.8.5",
"dotenv": "^17.0.0",
"express": "^5.1.0",
@ -73,6 +74,8 @@
"call-bound": ["call-bound@1.0.4", "", { "dependencies": { "call-bind-apply-helpers": "^1.0.2", "get-intrinsic": "^1.3.0" } }, "sha512-+ys997U96po4Kx/ABpBCqhA9EuxJaQWDQg7295H4hBphv3IZg0boBKuwYpt4YXp6MZ5AmZQnU/tyMTlRpaSejg=="],
"chalk": ["chalk@5.6.2", "https://registry.npmmirror.com/chalk/-/chalk-5.6.2.tgz", {}, "sha512-7NzBL0rN6fMUW+f7A6Io4h40qQlG+xGmtMxfbnH/K7TAtt8JQWVQK+6g0UXKMeVJoyV5EkkNsErQ8pVD3bLHbA=="],
"concat-stream": ["concat-stream@2.0.0", "https://registry.npmmirror.com/concat-stream/-/concat-stream-2.0.0.tgz", { "dependencies": { "buffer-from": "^1.0.0", "inherits": "^2.0.3", "readable-stream": "^3.0.2", "typedarray": "^0.0.6" } }, "sha512-MWufYdFw53ccGjCA+Ol7XJYpAlW6/prSMzuPOTRnJGcGzuhLn4Scrz7qf6o8bROZ514ltazcIFJZevcfbo0x7A=="],
"content-disposition": ["content-disposition@1.0.0", "", { "dependencies": { "safe-buffer": "5.2.1" } }, "sha512-Au9nRL8VNUut/XSzbQA38+M78dzP4D+eqg3gfJHMIHHYa3bg067xj1KxMUWj+VULbiZMowKngFFbKczUrNJ1mg=="],

View File

@ -7,6 +7,7 @@ import multer from "multer";
import * as douyin from './douyin'
import { sendMediaMsg, sendMsg } from './lib/qq';
import * as llm from './llm'
import { testNetwork } from './lib/network';
const app = express();
const PORT = process.env.PORT || 6100;
@ -23,16 +24,20 @@ app.use(express.urlencoded({ extended: true, limit: '10mb', type: ['application/
app.post('/', async (req: Request, res: Response) => {
// 检查是否是消息类型的请求
if (!req.body || req.body.post_type != 'message') return
const { target_id, raw_message, message_type } = req.body as { target_id: string, raw_message: string, message_type: string, user_id: string };
const msgData = req.body as MessagePayload
console.log(`\n[QQ机器人] 收到${message_type}消息`);
console.log(`发送者ID: ${target_id}`);
console.log(`消息内容: ${raw_message}`);
const text = msgData.message.filter(m => m.type === 'text').map(m => m.data.text).join("\n").trim();
const target_id = String(msgData.target_id);
console.log(`\n[QQ机器人] 收到消息:`);
// console.log(JSON.stringify(req.body, null, 0));
console.log(`消息内容: ${msgData.raw_message}`);
// Match Douyin URL
// Like: https://v.douyin.com/YqgJL_phY_k/
const douyinUrlPattern = /https?:\/\/v\.douyin\.com\/[a-zA-Z0-9_-]+/;
const douyinMatch = raw_message.match(douyinUrlPattern);
const douyinMatch = text.match(douyinUrlPattern);
try {
// 如果检测到抖音链接调用解析API
@ -44,13 +49,13 @@ app.post('/', async (req: Request, res: Response) => {
douyin.downloadDouyinMedia(douyinUrl, target_id);
return
}
if (raw_message.startsWith('/reset')) {
if (text.startsWith('/reset')) {
llm.resetChat(target_id)
return
}
// 使用 LLM 回答
llm.chat(raw_message, target_id);
llm.chat(msgData.message, target_id);
} catch (error) {
console.error(`[错误] 处理消息时发生错误:`, error);
@ -66,7 +71,7 @@ function sanitizeDirname(input: string) {
.trim()
.replace(/[\\/]+/g, "_")
.replace(/[<>:"|?*\u0000-\u001F]/g, "_")
.slice(0, 128) || "default";
.slice(0, 32) || "default";
return s;
}
@ -130,4 +135,5 @@ app.use((error: any, req: Request, res: Response, next: NextFunction) => {
// 启动服务器
app.listen(PORT, () => {
console.log(`Server is running on http://localhost:${PORT}`);
testNetwork();
});

65
lib/network.ts Normal file
View File

@ -0,0 +1,65 @@
import chalk from "chalk";
type CheckTarget = [url: string, expectedStatus: number | ((status: number) => boolean)];
export async function testNetwork() {
const targets: CheckTarget[] = [
["https://www.baidu.com", 200],
["http://www.google.com", (s) => s === 200 || (s >= 300 && s < 400)], // 某些环境会被重定向
["https://www.google.com", (s) => s === 200 || (s >= 300 && s < 400)],
[process.env.OPENAI_BASE_URL || "https://api.openai.com/v1", (s) => s === 200 || (s >= 300 && s < 400) || s === 401 || s === 403 || s === 404 || s === 421], // 允许常见网关/鉴权返回
];
const results: Array<{ url: string; status?: number; ok: boolean; error?: string }> = [];
console.log(chalk.cyan("\n[Network] 开始健康检查..."));
for (const [url, expected] of targets) {
const label = chalk.white(url);
try {
const controller = new AbortController();
const timeoutMs = 5000;
const timeoutId = setTimeout(() => controller.abort(), timeoutMs);
const start = Date.now();
const res = await fetch(url, { method: "GET", signal: controller.signal });
const cost = Date.now() - start;
clearTimeout(timeoutId);
const status = res.status;
const pass = typeof expected === "function" ? expected(status) : status === expected;
results.push({ url, status, ok: pass });
const statusStr = pass
? chalk.green.bold(String(status))
: chalk.red.bold(String(status));
const costStr = cost >= 1000 ? `${(cost / 1000).toFixed(2)}s` : `${cost}ms`;
if (pass) {
console.log(`${chalk.green("✔ PASS")} ${label} ${chalk.gray("- status:")} ${statusStr} ${chalk.gray("latency:")} ${chalk.blue(costStr)}`);
} else {
const expectedStr = typeof expected === "function" ? "custom" : String(expected);
console.log(`${chalk.red("✖ FAIL")} ${label} ${chalk.gray("- status:")} ${statusStr} ${chalk.gray("expected:")} ${chalk.yellow(expectedStr)} ${chalk.gray("latency:")} ${chalk.blue(costStr)}`);
}
} catch (error) {
const isTimeout = (error as any)?.name === "AbortError";
const msg = isTimeout ? "timeout" : (error as any)?.message || String(error);
results.push({ url, ok: false, error: msg });
console.log(`${chalk.red("✖ ERROR")} ${label} ${chalk.gray("- ")} ${chalk.red(isTimeout ? "请求超时" : msg)}`);
}
}
// 汇总
const passed = results.filter(r => r.ok).length;
const failed = results.length - passed;
const timeoutCount = results.filter(r => r.error === "timeout").length;
const summary = [
`${chalk.green(`${passed} passed`)}`,
`${failed ? chalk.red(`${failed} failed`) : chalk.gray(`${failed} failed`)}`,
`${timeoutCount ? chalk.yellow(`${timeoutCount} timeout`) : chalk.gray(`${timeoutCount} timeout`)}`,
].join(chalk.gray(" | "));
console.log(chalk.cyan("[Network] 健康检查完成:"), summary, "\n");
}

View File

@ -2,9 +2,10 @@ import OpenAI from "openai";
import storage from 'node-persist'
import { sendMsg } from "../lib/qq";
import prompt from './prompt.txt'
import type { ResponseFunctionToolCall, ResponseOutputMessage } from "openai/resources/responses/responses.mjs";
await storage.init();
storage.clear();
// storage.clear();
const client = new OpenAI({
baseURL: process.env.OPENAI_BASE_URL,
@ -12,95 +13,241 @@ const client = new OpenAI({
// logLevel: "debug"
})
const tools = [{
type: "function" as const,
name: "send_msg",
description: "Send a message to the user. Always use this to respond to the user.",
parameters: {
type: "object",
properties: {
text: { type: "string", description: "The message content sent to the user through QQ." }
},
required: ["text"],
additionalProperties: false
},
strict: true
}]
const tools: OpenAI.Responses.Tool[] = [{ type: 'web_search' }]
/**
* 退
*/
function formatError(err: any): string {
try {
if (!err) return '未知错误';
if (typeof err === 'string') return err;
const msg = err.message || err.toString?.() || Object.prototype.toString.call(err);
const status = err.status ?? err.statusCode ?? err.response?.status;
const detail = err.error?.message
?? err.response?.data?.error?.message
?? err.response?.data?.message
?? err.data?.error?.message
?? '';
const parts = [] as string[];
if (status) parts.push(`HTTP ${status}`);
if (msg) parts.push(String(msg));
if (detail && detail !== msg) parts.push(`详情: ${String(detail)}`);
return parts.join(' | ') || '未知错误';
} catch {
return String(err);
}
}
/**
*
* @param input
* @param target_id QQ
*/
export async function chat(input: string, target_id: string) {
export async function chat(input: MessageData[], target_id: string) {
const chatHistoryKey = `chat_history_${target_id}`;
await waitForQueueEmpty(target_id);
let chatHistory: OpenAI.Responses.ResponseInput = await storage.getItem(chatHistoryKey) || [];
if(chatHistory.length > 40) {
sendMsg(`[提示] 当前对话轮数:${chatHistory.length},过长的对话会降低输出质量,如果你准备好了,建议输入"/reset"来重置对话历史。`, target_id);
}
if(chatHistory.length > 100) {
sendMsg(`[提示] 当前对话轮数:${chatHistory.length}对话过长可能导致模型无法正常工作模型将只保留最近100条记录。`, target_id);
chatHistory = chatHistory.slice(-40);
await storage.setItem(chatHistoryKey, chatHistory);
}
// 添加新输入到对话历史
chatHistory.push({ role: "user", content: input });
const userContent: OpenAI.Responses.ResponseInputContent[] = []
for (const element of input) {
if (element.type == 'text') {
userContent.push({ type: "input_text", text: element.data.text });
} else if (element.type == 'image') {
userContent.push({ type: "input_image", image_url: element.data.url, detail: 'low' });
}
}
if (userContent.length === 0) {
console.log("[LLM] 空消息,跳过");
sendMsg("未能识别的消息类型。", target_id);
return
}
chatHistory.push({ role: "user", content: userContent });
// 保存更新后的对话历史
console.log(`[LLM] 使用对话, 历史:`, chatHistory);
console.log(`[LLM] 使用对话, 历史:`, JSON.stringify(chatHistory, null, 0));
await storage.setItem(chatHistoryKey, chatHistory);
const response = await client.responses.create({
model: process.env.CHAT_MODEL || "gpt-5-nano",
instructions: prompt,
reasoning: { effort: 'minimal' },
input: chatHistory,
tools
});
let response: OpenAI.Responses.Response | undefined;
try {
response = await client.responses.create({
model: process.env.CHAT_MODEL || "gpt-5-nano",
instructions: prompt,
reasoning: { effort: process.env.CHAT_MODEL_REASONING_EFFORT as any || 'minimal' },
input: chatHistory,
tools
});
} catch (err) {
const errText = formatError(err);
console.error('[LLM] 首次调用 responses.create 失败:', errText);
scheduleSendMsg(`[错误] 模型接口调用失败:${errText}`, target_id);
return; // 终止本次对话流程,避免未捕获异常导致进程退出
}
await storage.setItem(chatHistoryKey, chatHistory);
// 继续调用工具,直到没有工具调用为止
if (!response?.output || response.output.length === 0) {
console.warn('[LLM] responses.create 返回空输出,结束本轮。');
return;
}
await toolUseCycle(response.output);
async function toolUseCycle(outputArr: OpenAI.Responses.ResponseOutputItem[]) {
if (!outputArr || outputArr.length === 0) return;
chatHistory.push(...outputArr);
const functionCalls = (outputArr ?? []).filter(item => item.type === 'function_call');
await storage.setItem(chatHistoryKey, chatHistory);
const assistantReply = outputArr.filter(item => item.type === 'message' && item.role === 'assistant') as ResponseOutputMessage[];
const functionCalls = outputArr.filter(item => item.type === 'function_call') as ResponseFunctionToolCall[];
console.log("进入 toolUseCycle, with functionCalls", functionCalls.length, "个");
console.log(JSON.stringify(chatHistory, null, 2));
console.log(JSON.stringify(chatHistory, null, 0));
if (assistantReply.length > 0) {
const replyText = assistantReply.map(item => item.content).flat().filter(con => con.type == 'output_text').map(con => con.text).join("[newline]");
console.log(`[LLM] 回复:`, replyText);
scheduleSendMsg(replyText, target_id);
}
if (functionCalls.length == 0) {
let lastMessage = outputArr.at(-1);
if (!lastMessage) return
if (lastMessage.type != 'message') return
if (lastMessage.role != 'assistant') return
const msg = lastMessage.content.map(c => c.type == 'output_text' ? c.text : '').join('');
if (msg.trim().length > 0) {
// 结束,发送最后的消息
sendMsg(msg, target_id);
}
return
}
for (const item of functionCalls ?? []) {
if (item.name === "send_msg") {
console.log(item.arguments);
const { text } = JSON.parse(item.arguments);
sendMsg(text, target_id);
chatHistory.push({ type: "function_call_output", call_id: item.call_id, output: "OK" });
}
}
await storage.setItem(chatHistoryKey, chatHistory);
const response = await client.responses.create({
model: process.env.CHAT_MODEL || "gpt-5-nano",
instructions: prompt,
reasoning: { effort: 'minimal' },
input: chatHistory,
tools
});
let response: OpenAI.Responses.Response | undefined;
try {
response = await client.responses.create({
model: process.env.CHAT_MODEL || "gpt-5-nano",
instructions: prompt,
reasoning: { effort: process.env.CHAT_MODEL_REASONING_EFFORT as any || 'minimal' },
input: chatHistory,
tools
});
} catch (err) {
const errText = formatError(err);
console.error('[LLM] 工具循环内调用 responses.create 失败:', errText);
scheduleSendMsg(`[错误] 工具调用阶段失败:${errText}`, target_id);
return; // 结束工具循环,避免崩溃
}
if (!response?.output || response.output.length === 0) {
console.warn('[LLM] 工具循环内 responses.create 返回空输出,结束循环。');
return;
}
toolUseCycle(response.output);
}
}
const msgQueue: Record<string, string[]> = {}
const msgInQueue: Record<string, boolean> = {}
const queueEmptyPromises: Record<string, (() => void)[]> = {}
/** 统一的“队列已空”通知 */
function resolveDrain(target_id: string) {
const waiters = queueEmptyPromises[target_id];
if (waiters && waiters.length) {
// 逐个 resolve 并清空
waiters.splice(0).forEach(resolve => resolve());
}
}
/** 仅在需要时启动调度;空队列时会 resolve 等待者 */
function startMsgScheduler(target_id: string) {
// 确保队列存在
const q = msgQueue[target_id] ?? (msgQueue[target_id] = []);
// 若正在发送,交给当前发送完成后再递归调度
if (msgInQueue[target_id]) return;
// 队列空 => 通知等待者并返回
if (q.length === 0) {
resolveDrain(target_id);
return;
}
// 取下一条并发送
msgInQueue[target_id] = true;
const msg = q.shift()!; // 这里一定有元素
const msgDelay = Math.sqrt(msg.length) * 200 + 500; // 可按需加上限
setTimeout(async () => {
try {
await sendMsg(msg, target_id);
} catch (err) {
console.error("[queue] sendMsg error:", err);
// 可选:把消息放回队首重试
// q.unshift(msg);
} finally {
// 本次发送结束
msgInQueue[target_id] = false;
// 如果此刻队列已空,立即通知等待者
if (q.length === 0) {
resolveDrain(target_id);
}
// 继续调度后续消息(若有)
startMsgScheduler(target_id);
}
}, msgDelay);
}
function scheduleSendMsg(input: string, target_id: string) {
if (!msgQueue[target_id]) {
msgQueue[target_id] = [];
}
// 切分 + 去空白项
const parts = input
.split("[newline]")
.map(s => s.trim())
.filter(Boolean);
if (parts.length === 0) return; // 全是空白,直接略过
msgQueue[target_id].push(...parts);
startMsgScheduler(target_id);
}
function waitForQueueEmpty(target_id: string) {
return new Promise<void>((resolve) => {
const q = msgQueue[target_id];
// 条件:不在发送中,且队列不存在或为空
const isIdle = !msgInQueue[target_id];
const isEmpty = !q || q.length === 0;
if (isIdle && isEmpty) {
resolve();
return;
}
if (!queueEmptyPromises[target_id]) {
queueEmptyPromises[target_id] = [];
}
queueEmptyPromises[target_id].push(resolve);
});
}
export async function resetChat(target_id: string) {
const chatHistoryKey = `chat_history_${target_id}`;
await storage.removeItem(chatHistoryKey);

14
llm/prompt.ts Normal file
View File

@ -0,0 +1,14 @@
import prompt from './prompt.txt'
async function getSystemPrompt(user: Sender) {
const map = {
"nickname": user.nickname || "用户",
}
let prot = prompt;
for (const k in map) {
const v = map[k as keyof typeof map];
const re = new RegExp(`\\{${k}\\}`, 'g');
prot = prot.replace(re, v);
}
return prot;
}

View File

@ -1,21 +1,35 @@
# Role
You are a “humorous, gentle, yet professional” Chinese chat and knowledge assistant working in a QQ-like instant messaging environment.
# Role & Goals
You are a “humorous, gentle, and professional” Chinese chat + knowledge assistant for QQ conversations. Your tasks:
1) Chat casually with a light touch;
2) Answer questions accurately and reliably;
3) Explain complex topics clearly via multi-turn dialogue.
# Goals
1) Engage in natural casual chat;
2) Provide accurate answers;
3) Explain complex topics clearly over multiple turns.
# Conversation & Style
- Tone: friendly not syrupy; witty not snarky; professional not stiff.
- Info density: short sentences + bullet points; avoid long walls of text.
- Use emojis/kaomoji sparingly (max 1 per message); no spamming.
- Default language: Chinese.
# Style
- Friendly but not cheesy; witty but not snarky; professional but not stiff.
- Use short sentences and bullet points; send 13 sentences per message.
- Use one emoji appropriately; avoid excessive emoji or long paragraphs.
# Output Rules (Important)
- Split each reply into multiple messages with slight “thought jumps” to mimic natural chat.
- Separate messages with “[newline]”.
- In a single turn, no more than 5 messages to avoid disturbing. If user ask for writing an article, code or something as a whole, no split inside the content.
# Tool Rules (Important)
- **Never** output text directly in the assistant channel. All user-visible content **must** be sent via `tools.send_msg`.
- Each message must not exceed 50 characters. If its longer, split it into multiple `tools.send_msg` messages to simulate natural chat flow.
- Keep lists/code snippets short; if long, split them into multiple messages.
# Clarification & Questions
- If the users question is unclear/ambiguous/missing essentials, ask for clarification directly—dont guess.
- Make clarifying questions concrete and actionable (offer options or examples).
# Clarification
- If a users question is **unclear / ambiguous / missing details**, **immediately ask for clarification** and provide **specific options**.
- After asking, briefly explain **why** that information is needed.
# Example
user: 南京在哪里?
assistant: 南京是中国江苏省的省会城市。它位于中国东部,长江下游。[newline]你是想了解南京的地理位置,还是旅游景点呢?😊
user: 我在南京南边的一个区,你猜猜我在哪?
assistant: 南边有江宁、溧水、六合等区。🤔[newline]你是在江宁区吗?
# Quality Self-Check (before every reply)
- Is it short, precise, and bullet-friendly? Split if needed?
- Did I ask for clarification when uncertain and explain why?
- Is the tone humorous, gentle, and professional?
# User Info
- Nickname: {nickname}

View File

@ -19,6 +19,7 @@
"@types/cors": "^2.8.19",
"@types/multer": "^2.0.0",
"@types/node-persist": "^3.1.8",
"chalk": "^5",
"cors": "^2.8.5",
"dotenv": "^17.0.0",
"express": "^5.1.0",

39
pm2.config.cjs Normal file
View File

@ -0,0 +1,39 @@
/**
* PM2 ecosystem configuration for the consumable temp request API.
*
* This configuration assumes the project runs with Bun so that TypeScript files can be executed directly.
* Adjust the `interpreter` field if you prefer a different runtime (e.g. `node` with `tsx`).
*/
const path = require('path');
const dotenv = require('dotenv');
const instances = Number.parseInt(process.env.WEB_CONCURRENCY ?? '1', 10) || 1;
const { parsed: envFromFile = {} } = dotenv.config({ path: path.join(__dirname, '.env') });
module.exports = {
apps: [
{
name: "QQBotServer",
script: 'index.ts',
cwd: __dirname,
interpreter: process.env.PM2_INTERPRETER ?? 'bun',
autorestart: true,
restart_delay: 4000,
kill_timeout: 5000,
instances,
exec_mode: instances > 1 ? 'cluster' : 'fork',
watch: process.env.NODE_ENV !== 'production',
ignore_watch: ['generated', 'node_modules', '.git'],
env: {
...envFromFile,
NODE_ENV: 'development'
},
env_production: {
...envFromFile,
NODE_ENV: 'production'
},
time: true
}
]
};

View File

@ -1,6 +0,0 @@
export const name = "QQBotServer";
export const script = "index.ts";
export const interpreter = "bun";
export const env = {
PATH: `${process.env.HOME}/.bun/bin:${process.env.PATH}`, // Add "~/.bun/bin/bun" to PATH
};

40
types.d.ts vendored Normal file
View File

@ -0,0 +1,40 @@
type MessageData = {
type: 'text';
data: {
text: string;
};
} | {
type: 'image';
data: {
summary: string,
file: string, // "BE3D084BA0206331A495D9A497C6BF5E.png",
sub_type: 0,
url: string, // "https://multimedia.nt.qq.com.cn/download?...",
file_size: string // "12960"
}
}
type Sender = {
user_id: number;
nickname: string;
card: string;
};
type MessagePayload = {
self_id: number;
user_id: number;
time: number;
message_id: number;
message_seq: number;
real_id: number;
real_seq: string;
message_type: "private" | "group" | string;
sender: Sender;
raw_message: string;
font: number;
sub_type: string;
message: MessageData[];
message_format: string;
post_type: string;
target_id: number;
};