mirror of
https://github.com/ikechan8370/chatgpt-plugin.git
synced 2025-12-17 22:07:10 +00:00
Merge branch 'v2' of https://github.com/ikechan8370/chatgpt-plugin into v2
This commit is contained in:
commit
e0054521d3
18 changed files with 1749 additions and 117 deletions
|
|
@ -41,6 +41,8 @@
|
||||||
* 2023-05-29 支持gpt-4 API.必应无需cookie即可对话(Sydney和自定义模式)
|
* 2023-05-29 支持gpt-4 API.必应无需cookie即可对话(Sydney和自定义模式)
|
||||||
* 2023-07 支持智能模式,机器人可以实现禁言、群名片/头衔(需给机器人管理员/群主)、分享音乐视频、主动发音频、对接ap,sr和喵喵等插件、联网搜索等,需api模式0613系列模型。智能模式所需的额外api和搜索api分别可以参考[chatgpt-plugin-extras](https://github.com/ikechan8370/chatgpt-plugin-extras) 和 [search-api](https://github.com/ikechan8370/search-api) 自行搭建,其中后者提供了一个公益版本,前者可使用[huggingface](https://huggingface.co/spaces/ikechan8370/cp-extra)部署
|
* 2023-07 支持智能模式,机器人可以实现禁言、群名片/头衔(需给机器人管理员/群主)、分享音乐视频、主动发音频、对接ap,sr和喵喵等插件、联网搜索等,需api模式0613系列模型。智能模式所需的额外api和搜索api分别可以参考[chatgpt-plugin-extras](https://github.com/ikechan8370/chatgpt-plugin-extras) 和 [search-api](https://github.com/ikechan8370/search-api) 自行搭建,其中后者提供了一个公益版本,前者可使用[huggingface](https://huggingface.co/spaces/ikechan8370/cp-extra)部署
|
||||||
* 2023-09-10 支持来自claude.ai的claude-2模型
|
* 2023-09-10 支持来自claude.ai的claude-2模型
|
||||||
|
* 2023-10-19 支持读取文件,(目前适配必应模式和Claude2模式)
|
||||||
|
* 2023-10-25 增加支持通义千问官方API
|
||||||
### 如果觉得这个插件有趣或者对你有帮助,请点一个star吧!
|
### 如果觉得这个插件有趣或者对你有帮助,请点一个star吧!
|
||||||
|
|
||||||
## 版本要求
|
## 版本要求
|
||||||
|
|
|
||||||
280
apps/chat.js
280
apps/chat.js
|
|
@ -1,77 +1,82 @@
|
||||||
import plugin from '../../../lib/plugins/plugin.js'
|
import plugin from '../../../lib/plugins/plugin.js'
|
||||||
import _ from 'lodash'
|
import _ from 'lodash'
|
||||||
import { Config, defaultOpenAIAPI } from '../utils/config.js'
|
import {Config, defaultOpenAIAPI} from '../utils/config.js'
|
||||||
import { v4 as uuid } from 'uuid'
|
import {v4 as uuid} from 'uuid'
|
||||||
import delay from 'delay'
|
import delay from 'delay'
|
||||||
import { ChatGPTAPI } from '../utils/openai/chatgpt-api.js'
|
import {ChatGPTAPI} from '../utils/openai/chatgpt-api.js'
|
||||||
import { BingAIClient } from '@waylaidwanderer/chatgpt-api'
|
import {BingAIClient} from '@waylaidwanderer/chatgpt-api'
|
||||||
import SydneyAIClient from '../utils/SydneyAIClient.js'
|
import SydneyAIClient from '../utils/SydneyAIClient.js'
|
||||||
import { PoeClient } from '../utils/poe/index.js'
|
import {PoeClient} from '../utils/poe/index.js'
|
||||||
import AzureTTS from '../utils/tts/microsoft-azure.js'
|
import AzureTTS from '../utils/tts/microsoft-azure.js'
|
||||||
import VoiceVoxTTS from '../utils/tts/voicevox.js'
|
import VoiceVoxTTS from '../utils/tts/voicevox.js'
|
||||||
import Version from '../utils/version.js'
|
import Version from '../utils/version.js'
|
||||||
import {
|
import {
|
||||||
render,
|
|
||||||
renderUrl,
|
|
||||||
getMessageById,
|
|
||||||
makeForwardMsg,
|
|
||||||
upsertMessage,
|
|
||||||
randomString,
|
|
||||||
completeJSON,
|
completeJSON,
|
||||||
isImage,
|
extractContentFromFile,
|
||||||
getUserData,
|
formatDate,
|
||||||
|
formatDate2,
|
||||||
|
generateAudio,
|
||||||
getDefaultReplySetting,
|
getDefaultReplySetting,
|
||||||
isCN,
|
|
||||||
getMasterQQ,
|
|
||||||
getUserReplySetting,
|
|
||||||
getImageOcrText,
|
getImageOcrText,
|
||||||
getImg,
|
getImg,
|
||||||
getMaxModelTokens, formatDate, generateAudio, formatDate2, mkdirs, getUin
|
getMasterQQ,
|
||||||
|
getMaxModelTokens,
|
||||||
|
getMessageById,
|
||||||
|
getUin,
|
||||||
|
getUserData,
|
||||||
|
getUserReplySetting,
|
||||||
|
isCN,
|
||||||
|
isImage,
|
||||||
|
makeForwardMsg,
|
||||||
|
randomString,
|
||||||
|
render,
|
||||||
|
renderUrl,
|
||||||
|
upsertMessage
|
||||||
} from '../utils/common.js'
|
} from '../utils/common.js'
|
||||||
import { ChatGPTPuppeteer } from '../utils/browser.js'
|
import {ChatGPTPuppeteer} from '../utils/browser.js'
|
||||||
import { KeyvFile } from 'keyv-file'
|
import {KeyvFile} from 'keyv-file'
|
||||||
import { OfficialChatGPTClient } from '../utils/message.js'
|
import {OfficialChatGPTClient} from '../utils/message.js'
|
||||||
import fetch from 'node-fetch'
|
import fetch from 'node-fetch'
|
||||||
import { deleteConversation, getConversations, getLatestMessageIdByConversationId } from '../utils/conversation.js'
|
import {deleteConversation, getConversations, getLatestMessageIdByConversationId} from '../utils/conversation.js'
|
||||||
import { convertSpeaker, speakers } from '../utils/tts.js'
|
import {convertSpeaker, speakers} from '../utils/tts.js'
|
||||||
import ChatGLMClient from '../utils/chatglm.js'
|
import ChatGLMClient from '../utils/chatglm.js'
|
||||||
import { convertFaces } from '../utils/face.js'
|
import {convertFaces} from '../utils/face.js'
|
||||||
import { SlackClaudeClient } from '../utils/slack/slackClient.js'
|
import {SlackClaudeClient} from '../utils/slack/slackClient.js'
|
||||||
import { getPromptByName } from '../utils/prompts.js'
|
import {getPromptByName} from '../utils/prompts.js'
|
||||||
import BingDrawClient from '../utils/BingDraw.js'
|
import BingDrawClient from '../utils/BingDraw.js'
|
||||||
import XinghuoClient from '../utils/xinghuo/xinghuo.js'
|
import XinghuoClient from '../utils/xinghuo/xinghuo.js'
|
||||||
import Bard from '../utils/bard.js'
|
import Bard from '../utils/bard.js'
|
||||||
import { JinyanTool } from '../utils/tools/JinyanTool.js'
|
import {JinyanTool} from '../utils/tools/JinyanTool.js'
|
||||||
import { SendVideoTool } from '../utils/tools/SendBilibiliTool.js'
|
import {SendVideoTool} from '../utils/tools/SendBilibiliTool.js'
|
||||||
import { KickOutTool } from '../utils/tools/KickOutTool.js'
|
import {KickOutTool} from '../utils/tools/KickOutTool.js'
|
||||||
import { EditCardTool } from '../utils/tools/EditCardTool.js'
|
import {EditCardTool} from '../utils/tools/EditCardTool.js'
|
||||||
import { SearchVideoTool } from '../utils/tools/SearchBilibiliTool.js'
|
import {SearchVideoTool} from '../utils/tools/SearchBilibiliTool.js'
|
||||||
import { SearchMusicTool } from '../utils/tools/SearchMusicTool.js'
|
import {SearchMusicTool} from '../utils/tools/SearchMusicTool.js'
|
||||||
import { QueryStarRailTool } from '../utils/tools/QueryStarRailTool.js'
|
import {QueryStarRailTool} from '../utils/tools/QueryStarRailTool.js'
|
||||||
import { WebsiteTool } from '../utils/tools/WebsiteTool.js'
|
import {WebsiteTool} from '../utils/tools/WebsiteTool.js'
|
||||||
import { WeatherTool } from '../utils/tools/WeatherTool.js'
|
import {WeatherTool} from '../utils/tools/WeatherTool.js'
|
||||||
import { SerpTool } from '../utils/tools/SerpTool.js'
|
import {SerpTool} from '../utils/tools/SerpTool.js'
|
||||||
import { SerpIkechan8370Tool } from '../utils/tools/SerpIkechan8370Tool.js'
|
import {SerpIkechan8370Tool} from '../utils/tools/SerpIkechan8370Tool.js'
|
||||||
import { SendPictureTool } from '../utils/tools/SendPictureTool.js'
|
import {SendPictureTool} from '../utils/tools/SendPictureTool.js'
|
||||||
import { SerpImageTool } from '../utils/tools/SearchImageTool.js'
|
import {SerpImageTool} from '../utils/tools/SearchImageTool.js'
|
||||||
import { ImageCaptionTool } from '../utils/tools/ImageCaptionTool.js'
|
import {ImageCaptionTool} from '../utils/tools/ImageCaptionTool.js'
|
||||||
import { SendAudioMessageTool } from '../utils/tools/SendAudioMessageTool.js'
|
import {SendAudioMessageTool} from '../utils/tools/SendAudioMessageTool.js'
|
||||||
import { ProcessPictureTool } from '../utils/tools/ProcessPictureTool.js'
|
import {ProcessPictureTool} from '../utils/tools/ProcessPictureTool.js'
|
||||||
import { APTool } from '../utils/tools/APTool.js'
|
import {APTool} from '../utils/tools/APTool.js'
|
||||||
import { QueryGenshinTool } from '../utils/tools/QueryGenshinTool.js'
|
import {QueryGenshinTool} from '../utils/tools/QueryGenshinTool.js'
|
||||||
import { HandleMessageMsgTool } from '../utils/tools/HandleMessageMsgTool.js'
|
import {HandleMessageMsgTool} from '../utils/tools/HandleMessageMsgTool.js'
|
||||||
import { QueryUserinfoTool } from '../utils/tools/QueryUserinfoTool.js'
|
import {QueryUserinfoTool} from '../utils/tools/QueryUserinfoTool.js'
|
||||||
import { EliMovieTool } from '../utils/tools/EliMovieTool.js'
|
import {EliMovieTool} from '../utils/tools/EliMovieTool.js'
|
||||||
import { EliMusicTool } from '../utils/tools/EliMusicTool.js'
|
import {EliMusicTool} from '../utils/tools/EliMusicTool.js'
|
||||||
import { SendMusicTool } from '../utils/tools/SendMusicTool.js'
|
import {SendMusicTool} from '../utils/tools/SendMusicTool.js'
|
||||||
import { SendDiceTool } from '../utils/tools/SendDiceTool.js'
|
import {SendDiceTool} from '../utils/tools/SendDiceTool.js'
|
||||||
import { SendAvatarTool } from '../utils/tools/SendAvatarTool.js'
|
import {SendAvatarTool} from '../utils/tools/SendAvatarTool.js'
|
||||||
import { SendMessageToSpecificGroupOrUserTool } from '../utils/tools/SendMessageToSpecificGroupOrUserTool.js'
|
import {SendMessageToSpecificGroupOrUserTool} from '../utils/tools/SendMessageToSpecificGroupOrUserTool.js'
|
||||||
import { SetTitleTool } from '../utils/tools/SetTitleTool.js'
|
import {SetTitleTool} from '../utils/tools/SetTitleTool.js'
|
||||||
import { solveCaptchaOneShot } from '../utils/bingCaptcha.js'
|
import {solveCaptchaOneShot} from '../utils/bingCaptcha.js'
|
||||||
import { ClaudeAIClient } from '../utils/claude.ai/index.js'
|
import {ClaudeAIClient} from '../utils/claude.ai/index.js'
|
||||||
import fs from 'fs'
|
import {getProxy} from '../utils/proxy.js'
|
||||||
import { getProxy } from '../utils/proxy.js'
|
import {QwenApi} from '../utils/alibaba/qwen-api.js'
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await import('@azure/openai')
|
await import('@azure/openai')
|
||||||
|
|
@ -179,6 +184,12 @@ export class chatgpt extends plugin {
|
||||||
reg: '^#星火(搜索|查找)助手',
|
reg: '^#星火(搜索|查找)助手',
|
||||||
fnc: 'searchxhBot'
|
fnc: 'searchxhBot'
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
/** 命令正则匹配 */
|
||||||
|
reg: '^#qwen[sS]*',
|
||||||
|
/** 执行方法 */
|
||||||
|
fnc: 'qwen'
|
||||||
|
},
|
||||||
{
|
{
|
||||||
/** 命令正则匹配 */
|
/** 命令正则匹配 */
|
||||||
reg: toggleMode === 'at' ? '^[^#][sS]*' : '^#chat[^gpt][sS]*',
|
reg: toggleMode === 'at' ? '^[^#][sS]*' : '^#chat[^gpt][sS]*',
|
||||||
|
|
@ -366,6 +377,14 @@ export class chatgpt extends plugin {
|
||||||
await redis.del(`CHATGPT:CONVERSATIONS:${e.sender.user_id}`)
|
await redis.del(`CHATGPT:CONVERSATIONS:${e.sender.user_id}`)
|
||||||
await this.reply('已结束当前对话,请@我进行聊天以开启新的对话', true)
|
await this.reply('已结束当前对话,请@我进行聊天以开启新的对话', true)
|
||||||
}
|
}
|
||||||
|
} else if (use === 'qwen') {
|
||||||
|
let c = await redis.get(`CHATGPT:CONVERSATIONS_QWEN:${e.sender.user_id}`)
|
||||||
|
if (!c) {
|
||||||
|
await this.reply('当前没有开启对话', true)
|
||||||
|
} else {
|
||||||
|
await redis.del(`CHATGPT:CONVERSATIONS_QWEN:${e.sender.user_id}`)
|
||||||
|
await this.reply('已结束当前对话,请@我进行聊天以开启新的对话', true)
|
||||||
|
}
|
||||||
} else if (use === 'bing') {
|
} else if (use === 'bing') {
|
||||||
let c = await redis.get(`CHATGPT:CONVERSATIONS_BING:${e.sender.user_id}`)
|
let c = await redis.get(`CHATGPT:CONVERSATIONS_BING:${e.sender.user_id}`)
|
||||||
if (!c) {
|
if (!c) {
|
||||||
|
|
@ -427,6 +446,14 @@ export class chatgpt extends plugin {
|
||||||
await redis.del(`CHATGPT:CONVERSATIONS:${qq}`)
|
await redis.del(`CHATGPT:CONVERSATIONS:${qq}`)
|
||||||
await this.reply(`已结束${atUser}的对话,TA仍可以@我进行聊天以开启新的对话`, true)
|
await this.reply(`已结束${atUser}的对话,TA仍可以@我进行聊天以开启新的对话`, true)
|
||||||
}
|
}
|
||||||
|
} else if (use === 'qwen') {
|
||||||
|
let c = await redis.get(`CHATGPT:CONVERSATIONS_QWEN:${qq}`)
|
||||||
|
if (!c) {
|
||||||
|
await this.reply(`当前${atUser}没有开启对话`, true)
|
||||||
|
} else {
|
||||||
|
await redis.del(`CHATGPT:CONVERSATIONS_QWEN:${qq}`)
|
||||||
|
await this.reply(`已结束${atUser}的对话,TA仍可以@我进行聊天以开启新的对话`, true)
|
||||||
|
}
|
||||||
} else if (use === 'bing') {
|
} else if (use === 'bing') {
|
||||||
let c = await redis.get(`CHATGPT:CONVERSATIONS_BING:${qq}`)
|
let c = await redis.get(`CHATGPT:CONVERSATIONS_BING:${qq}`)
|
||||||
if (!c) {
|
if (!c) {
|
||||||
|
|
@ -1029,6 +1056,10 @@ export class chatgpt extends plugin {
|
||||||
key = `CHATGPT:CONVERSATIONS_AZURE:${e.sender.user_id}`
|
key = `CHATGPT:CONVERSATIONS_AZURE:${e.sender.user_id}`
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
case 'qwen': {
|
||||||
|
key = `CHATGPT:CONVERSATIONS_QWEN:${(e.isGroup && Config.groupMerge) ? e.group_id.toString() : e.sender.user_id}`
|
||||||
|
break
|
||||||
|
}
|
||||||
}
|
}
|
||||||
let ctime = new Date()
|
let ctime = new Date()
|
||||||
previousConversation = (key ? await redis.get(key) : null) || JSON.stringify({
|
previousConversation = (key ? await redis.get(key) : null) || JSON.stringify({
|
||||||
|
|
@ -1439,6 +1470,25 @@ export class chatgpt extends plugin {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async qwen (e) {
|
||||||
|
if (!Config.allowOtherMode) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
let ats = e.message.filter(m => m.type === 'at')
|
||||||
|
if (!(e.atme || e.atBot) && ats.length > 0) {
|
||||||
|
if (Config.debug) {
|
||||||
|
logger.mark('艾特别人了,没艾特我,忽略#xh')
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
let prompt = _.replace(e.raw_message.trimStart(), '#qwen', '').trim()
|
||||||
|
if (prompt.length === 0) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
await this.abstractChat(e, prompt, 'qwen')
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
async xh (e) {
|
async xh (e) {
|
||||||
if (!Config.allowOtherMode) {
|
if (!Config.allowOtherMode) {
|
||||||
return false
|
return false
|
||||||
|
|
@ -1626,6 +1676,20 @@ export class chatgpt extends plugin {
|
||||||
logger.warn('获取群聊聊天记录失败,本次对话不携带聊天记录', err)
|
logger.warn('获取群聊聊天记录失败,本次对话不携带聊天记录', err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
let toSummaryFileContent
|
||||||
|
try {
|
||||||
|
if (e.source) {
|
||||||
|
let msgs = e.isGroup ? await e.group.getChatHistory(e.source.seq, 1) : await e.friend.getChatHistory(e.source.time, 1)
|
||||||
|
let sourceMsg = msgs[0]
|
||||||
|
let fileMsgElem = sourceMsg.message.find(msg => msg.type === 'file')
|
||||||
|
if (fileMsgElem) {
|
||||||
|
toSummaryFileContent = await extractContentFromFile(fileMsgElem, e)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn('读取文件内容出错, 忽略文件内容', err)
|
||||||
|
}
|
||||||
|
opt.toSummaryFileContent = toSummaryFileContent
|
||||||
} else {
|
} else {
|
||||||
// 重新创建client,因为token可能换到别的了
|
// 重新创建client,因为token可能换到别的了
|
||||||
if (bingToken?.indexOf('=') > -1) {
|
if (bingToken?.indexOf('=') > -1) {
|
||||||
|
|
@ -1893,40 +1957,30 @@ export class chatgpt extends plugin {
|
||||||
debug: Config.debug,
|
debug: Config.debug,
|
||||||
proxy: Config.proxy
|
proxy: Config.proxy
|
||||||
})
|
})
|
||||||
let fileUrl, filename, attachments
|
let toSummaryFileContent
|
||||||
if (e.source && e.source.message === '[文件]') {
|
try {
|
||||||
if (e.isGroup) {
|
if (e.source) {
|
||||||
let source = (await e.group.getChatHistory(e.source.seq, 1))[0]
|
let msgs = e.isGroup ? await e.group.getChatHistory(e.source.seq, 1) : await e.friend.getChatHistory(e.source.time, 1)
|
||||||
let file = source.message.find(m => m.type === 'file')
|
let sourceMsg = msgs[0]
|
||||||
if (file) {
|
let fileMsgElem = sourceMsg.message.find(msg => msg.type === 'file')
|
||||||
filename = file.name
|
if (fileMsgElem) {
|
||||||
fileUrl = await e.group.getFileUrl(file.fid)
|
toSummaryFileContent = await extractContentFromFile(fileMsgElem, e)
|
||||||
}
|
|
||||||
} else {
|
|
||||||
let source = (await e.friend.getChatHistory(e.source.time, 1))[0]
|
|
||||||
let file = source.message.find(m => m.type === 'file')
|
|
||||||
if (file) {
|
|
||||||
filename = file.name
|
|
||||||
fileUrl = await e.group.getFileUrl(file.fid)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
} catch (err) {
|
||||||
|
logger.warn('读取文件内容出错, 忽略文件内容', err)
|
||||||
}
|
}
|
||||||
if (fileUrl) {
|
|
||||||
logger.info('文件地址:' + fileUrl)
|
let attachments = []
|
||||||
mkdirs('data/chatgpt/files')
|
if (toSummaryFileContent?.content) {
|
||||||
let destinationPath = 'data/chatgpt/files/' + filename
|
attachments.push({
|
||||||
const response = await fetch(fileUrl)
|
extracted_content: toSummaryFileContent.content,
|
||||||
const fileStream = fs.createWriteStream(destinationPath)
|
file_name: toSummaryFileContent.name,
|
||||||
await new Promise((resolve, reject) => {
|
file_type: 'pdf',
|
||||||
response.body.pipe(fileStream)
|
file_size: 200312,
|
||||||
response.body.on('error', (err) => {
|
totalPages: 20
|
||||||
reject(err)
|
|
||||||
})
|
|
||||||
fileStream.on('finish', () => {
|
|
||||||
resolve()
|
|
||||||
})
|
|
||||||
})
|
})
|
||||||
attachments = [await client.convertDocument(destinationPath, filename)]
|
logger.info(toSummaryFileContent.content)
|
||||||
}
|
}
|
||||||
if (conversationId) {
|
if (conversationId) {
|
||||||
return await client.sendMessage(prompt, conversationId, attachments)
|
return await client.sendMessage(prompt, conversationId, attachments)
|
||||||
|
|
@ -1976,6 +2030,57 @@ export class chatgpt extends plugin {
|
||||||
let completion = choices[0].message
|
let completion = choices[0].message
|
||||||
return { text: completion.content, message: completion }
|
return { text: completion.content, message: completion }
|
||||||
}
|
}
|
||||||
|
case 'qwen': {
|
||||||
|
let completionParams = {
|
||||||
|
parameters: {
|
||||||
|
top_p: Config.qwenTopP || 0.5,
|
||||||
|
top_k: Config.qwenTopK || 50,
|
||||||
|
seed: Config.qwenSeed > 0 ? Config.qwenSeed : Math.floor(Math.random() * 114514),
|
||||||
|
temperature: Config.qwenTemperature || 1,
|
||||||
|
enable_search: !!Config.qwenEnableSearch
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (Config.qwenModel) {
|
||||||
|
completionParams.model = Config.qwenModel
|
||||||
|
}
|
||||||
|
const currentDate = new Date().toISOString().split('T')[0]
|
||||||
|
async function um (message) {
|
||||||
|
return await upsertMessage(message, 'QWEN')
|
||||||
|
}
|
||||||
|
async function gm (id) {
|
||||||
|
return await getMessageById(id, 'QWEN')
|
||||||
|
}
|
||||||
|
let opts = {
|
||||||
|
apiKey: Config.qwenApiKey,
|
||||||
|
debug: false,
|
||||||
|
upsertMessage: um,
|
||||||
|
getMessageById: gm,
|
||||||
|
systemMessage: `You are ${Config.assistantLabel} ${useCast?.api || Config.promptPrefixOverride || defaultPropmtPrefix}
|
||||||
|
Current date: ${currentDate}`,
|
||||||
|
completionParams,
|
||||||
|
assistantLabel: Config.assistantLabel,
|
||||||
|
fetch: newFetch
|
||||||
|
}
|
||||||
|
this.qwenApi = new QwenApi(opts)
|
||||||
|
let option = {
|
||||||
|
timeoutMs: 600000,
|
||||||
|
completionParams
|
||||||
|
}
|
||||||
|
if (conversation) {
|
||||||
|
if (!conversation.conversationId) {
|
||||||
|
conversation.conversationId = uuid()
|
||||||
|
}
|
||||||
|
option = Object.assign(option, conversation)
|
||||||
|
}
|
||||||
|
let msg
|
||||||
|
try {
|
||||||
|
msg = await this.qwenApi.sendMessage(prompt, option)
|
||||||
|
} catch (err) {
|
||||||
|
logger.error(err)
|
||||||
|
throw new Error(err)
|
||||||
|
}
|
||||||
|
return msg
|
||||||
|
}
|
||||||
case 'bard': {
|
case 'bard': {
|
||||||
// 处理cookie
|
// 处理cookie
|
||||||
const matchesPSID = /__Secure-1PSID=([^;]+)/.exec(Config.bardPsid)
|
const matchesPSID = /__Secure-1PSID=([^;]+)/.exec(Config.bardPsid)
|
||||||
|
|
@ -2134,6 +2239,9 @@ export class chatgpt extends plugin {
|
||||||
}
|
}
|
||||||
option.systemMessage = system
|
option.systemMessage = system
|
||||||
if (conversation) {
|
if (conversation) {
|
||||||
|
if (!conversation.conversationId) {
|
||||||
|
conversation.conversationId = uuid()
|
||||||
|
}
|
||||||
option = Object.assign(option, conversation)
|
option = Object.assign(option, conversation)
|
||||||
}
|
}
|
||||||
if (Config.smartMode) {
|
if (Config.smartMode) {
|
||||||
|
|
|
||||||
|
|
@ -124,6 +124,11 @@ export class ChatgptManagement extends plugin {
|
||||||
fnc: 'useBardBasedSolution',
|
fnc: 'useBardBasedSolution',
|
||||||
permission: 'master'
|
permission: 'master'
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
reg: '^#chatgpt切换(通义千问|qwen|千问)$',
|
||||||
|
fnc: 'useQwenSolution',
|
||||||
|
permission: 'master'
|
||||||
|
},
|
||||||
{
|
{
|
||||||
reg: '^#chatgpt(必应|Bing)切换',
|
reg: '^#chatgpt(必应|Bing)切换',
|
||||||
fnc: 'changeBingTone',
|
fnc: 'changeBingTone',
|
||||||
|
|
@ -289,9 +294,9 @@ ${userSetting.useTTS === true ? '当前语音模式为' + Config.ttsMode : ''}`
|
||||||
const matchCommand = e.msg.match(/^#(chatgpt)?(vits|azure|vox)?语音(服务|角色列表)/)
|
const matchCommand = e.msg.match(/^#(chatgpt)?(vits|azure|vox)?语音(服务|角色列表)/)
|
||||||
if (matchCommand[3] === '服务') {
|
if (matchCommand[3] === '服务') {
|
||||||
await this.reply(`当前支持vox、vits、azure语音服务,可使用'#(vox|azure|vits)语音角色列表'查看支持的语音角色。
|
await this.reply(`当前支持vox、vits、azure语音服务,可使用'#(vox|azure|vits)语音角色列表'查看支持的语音角色。
|
||||||
|
|
||||||
vits语音:主要有赛马娘,原神中文,原神日语,崩坏 3 的音色、结果有随机性,语调可能很奇怪。
|
vits语音:主要有赛马娘,原神中文,原神日语,崩坏 3 的音色、结果有随机性,语调可能很奇怪。
|
||||||
|
|
||||||
vox语音:Voicevox 是一款由日本 DeNA 开发的语音合成软件,它可以将文本转换为自然流畅的语音。Voicevox 支持多种语言和声音,可以用于制作各种语音内容,如动画、游戏、广告等。Voicevox 还提供了丰富的调整选项,可以调整声音的音调、速度、音量等参数,以满足不同需求。除了桌面版软件外,Voicevox 还提供了 Web 版本和 API 接口,方便开发者在各种平台上使用。
|
vox语音:Voicevox 是一款由日本 DeNA 开发的语音合成软件,它可以将文本转换为自然流畅的语音。Voicevox 支持多种语言和声音,可以用于制作各种语音内容,如动画、游戏、广告等。Voicevox 还提供了丰富的调整选项,可以调整声音的音调、速度、音量等参数,以满足不同需求。除了桌面版软件外,Voicevox 还提供了 Web 版本和 API 接口,方便开发者在各种平台上使用。
|
||||||
|
|
||||||
azure语音:Azure 语音是微软 Azure 平台提供的一项语音服务,它可以帮助开发者将语音转换为文本、将文本转换为语音、实现自然语言理解和对话等功能。Azure 语音支持多种语言和声音,可以用于构建各种语音应用程序,如智能客服、语音助手、自动化电话系统等。Azure 语音还提供了丰富的 API 和 SDK,方便开发者在各种平台上集成使用。
|
azure语音:Azure 语音是微软 Azure 平台提供的一项语音服务,它可以帮助开发者将语音转换为文本、将文本转换为语音、实现自然语言理解和对话等功能。Azure 语音支持多种语言和声音,可以用于构建各种语音应用程序,如智能客服、语音助手、自动化电话系统等。Azure 语音还提供了丰富的 API 和 SDK,方便开发者在各种平台上集成使用。
|
||||||
|
|
@ -864,6 +869,7 @@ azure语音:Azure 语音是微软 Azure 平台提供的一项语音服务,
|
||||||
await this.reply('当前已经是星火模式了')
|
await this.reply('当前已经是星火模式了')
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async useAzureBasedSolution () {
|
async useAzureBasedSolution () {
|
||||||
let use = await redis.get('CHATGPT:USE')
|
let use = await redis.get('CHATGPT:USE')
|
||||||
if (use !== 'azure') {
|
if (use !== 'azure') {
|
||||||
|
|
@ -884,6 +890,16 @@ azure语音:Azure 语音是微软 Azure 平台提供的一项语音服务,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async useQwenSolution () {
|
||||||
|
let use = await redis.get('CHATGPT:USE')
|
||||||
|
if (use !== 'qwen') {
|
||||||
|
await redis.set('CHATGPT:USE', 'qwen')
|
||||||
|
await this.reply('已切换到基于通义千问的解决方案')
|
||||||
|
} else {
|
||||||
|
await this.reply('当前已经是通义千问模式了')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
async changeBingTone (e) {
|
async changeBingTone (e) {
|
||||||
let tongStyle = e.msg.replace(/^#chatgpt(必应|Bing)切换/, '')
|
let tongStyle = e.msg.replace(/^#chatgpt(必应|Bing)切换/, '')
|
||||||
if (!tongStyle) {
|
if (!tongStyle) {
|
||||||
|
|
@ -1275,9 +1291,9 @@ Poe 模式会调用 Poe 中的 Claude-instant 进行对话。需要提供 Cookie
|
||||||
const viewHost = Config.serverHost ? `http://${Config.serverHost}/` : `http://${await getPublicIP()}:${Config.serverPort || 3321}/`
|
const viewHost = Config.serverHost ? `http://${Config.serverHost}/` : `http://${await getPublicIP()}:${Config.serverPort || 3321}/`
|
||||||
const otp = randomString(6)
|
const otp = randomString(6)
|
||||||
await redis.set(
|
await redis.set(
|
||||||
`CHATGPT:SERVER_QUICK`,
|
'CHATGPT:SERVER_QUICK',
|
||||||
otp,
|
otp,
|
||||||
{ EX: 60000 }
|
{ EX: 60000 }
|
||||||
)
|
)
|
||||||
await this.reply(`请登录http://tools.alcedogroup.com/login?server=${viewHost}&otp=${otp}`, true)
|
await this.reply(`请登录http://tools.alcedogroup.com/login?server=${viewHost}&otp=${otp}`, true)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -741,6 +741,53 @@ export function supportGuoba () {
|
||||||
bottomHelpMessage: '开启后将通过反代访问bard',
|
bottomHelpMessage: '开启后将通过反代访问bard',
|
||||||
component: 'Switch'
|
component: 'Switch'
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
label: '以下为通义千问API方式的配置',
|
||||||
|
component: 'Divider'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenApiKey',
|
||||||
|
label: '通义千问API Key',
|
||||||
|
component: 'InputPassword'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenModel',
|
||||||
|
label: '通义千问模型',
|
||||||
|
bottomHelpMessage: '指明需要调用的模型,目前可选 qwen-turbo 和 qwen-plus',
|
||||||
|
component: 'Input'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenTopP',
|
||||||
|
label: '通义千问topP',
|
||||||
|
bottomHelpMessage: '生成时,核采样方法的概率阈值。例如,取值为0.8时,仅保留累计概率之和大于等于0.8的概率分布中的token,作为随机采样的候选集。取值范围为(0,1.0),取值越大,生成的随机性越高;取值越低,生成的随机性越低。默认值 0.5。注意,取值不要大于等于1',
|
||||||
|
component: 'InputNumber'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenTopK',
|
||||||
|
label: '通义千问topK',
|
||||||
|
bottomHelpMessage: '生成时,采样候选集的大小。例如,取值为50时,仅将单次生成中得分最高的50个token组成随机采样的候选集。取值越大,生成的随机性越高;取值越小,生成的确定性越高。注意:如果top_k的值大于100,top_k将采用默认值0,表示不启用top_k策略,此时仅有top_p策略生效。',
|
||||||
|
component: 'InputNumber'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenSeed',
|
||||||
|
label: '通义千问Seed',
|
||||||
|
bottomHelpMessage: '生成时,随机数的种子,用于控制模型生成的随机性。如果使用相同的种子,每次运行生成的结果都将相同;当需要复现模型的生成结果时,可以使用相同的种子。seed参数支持无符号64位整数类型。默认值 0, 表示每次随机生成',
|
||||||
|
component: 'InputNumber'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenTemperature',
|
||||||
|
label: '通义千问温度',
|
||||||
|
bottomHelpMessage: '用于控制随机性和多样性的程度。具体来说,temperature值控制了生成文本时对每个候选词的概率分布进行平滑的程度。较高的temperature值会降低概率分布的峰值,使得更多的低概率词被选择,生成结果更加多样化;而较低的temperature值则会增强概率分布的峰值,使得高概率词更容易被选择,生成结果更加确定。\n' +
|
||||||
|
'\n' +
|
||||||
|
'取值范围: (0, 2),系统默认值1.0',
|
||||||
|
component: 'InputNumber'
|
||||||
|
},
|
||||||
|
{
|
||||||
|
field: 'qwenEnableSearch',
|
||||||
|
label: '通义千问允许搜索',
|
||||||
|
bottomHelpMessage: '生成时,是否参考夸克搜索的结果。注意:打开搜索并不意味着一定会使用搜索结果;如果打开搜索,模型会将搜索结果作为prompt,进而“自行判断”是否生成结合搜索结果的文本,默认为false',
|
||||||
|
component: 'Switch'
|
||||||
|
},
|
||||||
{
|
{
|
||||||
label: '以下为杂七杂八的配置',
|
label: '以下为杂七杂八的配置',
|
||||||
component: 'Divider'
|
component: 'Divider'
|
||||||
|
|
|
||||||
|
|
@ -35,14 +35,18 @@
|
||||||
"ws": "^8.13.0"
|
"ws": "^8.13.0"
|
||||||
},
|
},
|
||||||
"optionalDependencies": {
|
"optionalDependencies": {
|
||||||
|
"xlsx": "^0.18.5",
|
||||||
|
"mammoth": "^1.6.0",
|
||||||
|
"pdfjs-dist": "^3.11.174",
|
||||||
|
"nodejs-pptx": "^1.2.4",
|
||||||
"@node-rs/jieba": "^1.6.2",
|
"@node-rs/jieba": "^1.6.2",
|
||||||
|
"cycletls": "^1.0.21",
|
||||||
"jimp": "^0.22.7",
|
"jimp": "^0.22.7",
|
||||||
"node-silk": "^0.1.0",
|
"node-silk": "^0.1.0",
|
||||||
"puppeteer-extra": "^3.3.6",
|
"puppeteer-extra": "^3.3.6",
|
||||||
"puppeteer-extra-plugin-recaptcha": "^3.6.8",
|
"puppeteer-extra-plugin-recaptcha": "^3.6.8",
|
||||||
"puppeteer-extra-plugin-stealth": "^2.11.2",
|
"puppeteer-extra-plugin-stealth": "^2.11.2",
|
||||||
"sharp": "^0.32.3",
|
"sharp": "^0.32.3"
|
||||||
"cycletls": "^1.0.21"
|
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"ts-node": "^10.9.1",
|
"ts-node": "^10.9.1",
|
||||||
|
|
|
||||||
|
|
@ -7,7 +7,7 @@ import fetch, {
|
||||||
import crypto from 'crypto'
|
import crypto from 'crypto'
|
||||||
import WebSocket from 'ws'
|
import WebSocket from 'ws'
|
||||||
import { Config, pureSydneyInstruction } from './config.js'
|
import { Config, pureSydneyInstruction } from './config.js'
|
||||||
import { formatDate, getMasterQQ, isCN, getUserData } from './common.js'
|
import { formatDate, getMasterQQ, isCN, getUserData, limitString } from './common.js'
|
||||||
import delay from 'delay'
|
import delay from 'delay'
|
||||||
import moment from 'moment'
|
import moment from 'moment'
|
||||||
import { getProxy } from './proxy.js'
|
import { getProxy } from './proxy.js'
|
||||||
|
|
@ -101,12 +101,12 @@ export default class SydneyAIClient {
|
||||||
this.opts.host = 'https://edgeservices.bing.com/edgesvc'
|
this.opts.host = 'https://edgeservices.bing.com/edgesvc'
|
||||||
}
|
}
|
||||||
logger.mark('使用host:' + this.opts.host)
|
logger.mark('使用host:' + this.opts.host)
|
||||||
let response = await fetch(`${this.opts.host}/turing/conversation/create?bundleVersion=1.1055.6`, fetchOptions)
|
let response = await fetch(`${this.opts.host}/turing/conversation/create?bundleVersion=1.1055.10`, fetchOptions)
|
||||||
let text = await response.text()
|
let text = await response.text()
|
||||||
let retry = 10
|
let retry = 10
|
||||||
while (retry >= 0 && response.status === 200 && !text) {
|
while (retry >= 0 && response.status === 200 && !text) {
|
||||||
await delay(400)
|
await delay(400)
|
||||||
response = await fetch(`${this.opts.host}/turing/conversation/create`, fetchOptions)
|
response = await fetch(`${this.opts.host}/turing/conversation/create?bundleVersion=1.1055.10`, fetchOptions)
|
||||||
text = await response.text()
|
text = await response.text()
|
||||||
retry--
|
retry--
|
||||||
}
|
}
|
||||||
|
|
@ -222,8 +222,8 @@ export default class SydneyAIClient {
|
||||||
timeout = Config.defaultTimeoutMs,
|
timeout = Config.defaultTimeoutMs,
|
||||||
firstMessageTimeout = Config.sydneyFirstMessageTimeout,
|
firstMessageTimeout = Config.sydneyFirstMessageTimeout,
|
||||||
groupId, nickname, qq, groupName, chats, botName, masterName,
|
groupId, nickname, qq, groupName, chats, botName, masterName,
|
||||||
messageType = 'Chat'
|
messageType = 'Chat',
|
||||||
|
toSummaryFileContent
|
||||||
} = opts
|
} = opts
|
||||||
// if (messageType === 'Chat') {
|
// if (messageType === 'Chat') {
|
||||||
// logger.warn('该Bing账户token已被限流,降级至使用非搜索模式。本次对话AI将无法使用Bing搜索返回的内容')
|
// logger.warn('该Bing账户token已被限流,降级至使用非搜索模式。本次对话AI将无法使用Bing搜索返回的内容')
|
||||||
|
|
@ -372,6 +372,10 @@ export default class SydneyAIClient {
|
||||||
let maxConv = Config.maxNumUserMessagesInConversation
|
let maxConv = Config.maxNumUserMessagesInConversation
|
||||||
const currentDate = moment().format('YYYY-MM-DDTHH:mm:ssZ')
|
const currentDate = moment().format('YYYY-MM-DDTHH:mm:ssZ')
|
||||||
const imageDate = await this.kblobImage(opts.imageUrl)
|
const imageDate = await this.kblobImage(opts.imageUrl)
|
||||||
|
if (toSummaryFileContent?.content) {
|
||||||
|
// message = `请不要进行搜索,用户的问题是:"${message}"`
|
||||||
|
messageType = 'Chat'
|
||||||
|
}
|
||||||
let argument0 = {
|
let argument0 = {
|
||||||
source: 'cib',
|
source: 'cib',
|
||||||
optionsSets,
|
optionsSets,
|
||||||
|
|
@ -415,10 +419,12 @@ export default class SydneyAIClient {
|
||||||
text: message,
|
text: message,
|
||||||
messageType,
|
messageType,
|
||||||
userIpAddress: await generateRandomIP(),
|
userIpAddress: await generateRandomIP(),
|
||||||
timestamp: currentDate
|
timestamp: currentDate,
|
||||||
|
privacy: 'Internal'
|
||||||
// messageType: 'SearchQuery'
|
// messageType: 'SearchQuery'
|
||||||
},
|
},
|
||||||
tone: 'Creative',
|
tone: 'Creative',
|
||||||
|
privacy: 'Internal',
|
||||||
conversationSignature,
|
conversationSignature,
|
||||||
participant: {
|
participant: {
|
||||||
id: clientId
|
id: clientId
|
||||||
|
|
@ -440,7 +446,7 @@ export default class SydneyAIClient {
|
||||||
}
|
}
|
||||||
// simulates document summary function on Edge's Bing sidebar
|
// simulates document summary function on Edge's Bing sidebar
|
||||||
// unknown character limit, at least up to 7k
|
// unknown character limit, at least up to 7k
|
||||||
if (groupId) {
|
if (groupId && !toSummaryFileContent?.content) {
|
||||||
context += '注意,你现在正在一个qq群里和人聊天,现在问你问题的人是' + `${nickname}(${qq})。`
|
context += '注意,你现在正在一个qq群里和人聊天,现在问你问题的人是' + `${nickname}(${qq})。`
|
||||||
if (Config.enforceMaster && master) {
|
if (Config.enforceMaster && master) {
|
||||||
if (qq === master) {
|
if (qq === master) {
|
||||||
|
|
@ -485,7 +491,19 @@ export default class SydneyAIClient {
|
||||||
}).join('\n')
|
}).join('\n')
|
||||||
context += '\n'
|
context += '\n'
|
||||||
}
|
}
|
||||||
if (context) {
|
if (toSummaryFileContent?.content) {
|
||||||
|
// 忽略context 不然可能会爆炸
|
||||||
|
obj.arguments[0].previousMessages.push({
|
||||||
|
author: 'user',
|
||||||
|
description: limitString(toSummaryFileContent?.content, 20000, true),
|
||||||
|
contextType: 'WebPage',
|
||||||
|
messageType: 'Context',
|
||||||
|
sourceName: toSummaryFileContent?.name,
|
||||||
|
sourceUrl: 'file:///C:/Users/turing/Downloads/Documents/' + toSummaryFileContent?.name || 'file.pdf',
|
||||||
|
// locale: 'und',
|
||||||
|
// privacy: 'Internal'
|
||||||
|
})
|
||||||
|
} else if (context) {
|
||||||
obj.arguments[0].previousMessages.push({
|
obj.arguments[0].previousMessages.push({
|
||||||
author: 'user',
|
author: 'user',
|
||||||
description: context,
|
description: context,
|
||||||
|
|
|
||||||
394
utils/alibaba/qwen-api.js
Normal file
394
utils/alibaba/qwen-api.js
Normal file
|
|
@ -0,0 +1,394 @@
|
||||||
|
var __assign = (this && this.__assign) || function () {
|
||||||
|
__assign = Object.assign || function(t) {
|
||||||
|
for (var s, i = 1, n = arguments.length; i < n; i++) {
|
||||||
|
s = arguments[i];
|
||||||
|
for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p))
|
||||||
|
t[p] = s[p];
|
||||||
|
}
|
||||||
|
return t;
|
||||||
|
};
|
||||||
|
return __assign.apply(this, arguments);
|
||||||
|
};
|
||||||
|
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
||||||
|
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
||||||
|
return new (P || (P = Promise))(function (resolve, reject) {
|
||||||
|
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
||||||
|
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
||||||
|
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
||||||
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
||||||
|
});
|
||||||
|
};
|
||||||
|
var __generator = (this && this.__generator) || function (thisArg, body) {
|
||||||
|
var _ = { label: 0, sent: function() { if (t[0] & 1) throw t[1]; return t[1]; }, trys: [], ops: [] }, f, y, t, g;
|
||||||
|
return g = { next: verb(0), "throw": verb(1), "return": verb(2) }, typeof Symbol === "function" && (g[Symbol.iterator] = function() { return this; }), g;
|
||||||
|
function verb(n) { return function (v) { return step([n, v]); }; }
|
||||||
|
function step(op) {
|
||||||
|
if (f) throw new TypeError("Generator is already executing.");
|
||||||
|
while (g && (g = 0, op[0] && (_ = 0)), _) try {
|
||||||
|
if (f = 1, y && (t = op[0] & 2 ? y["return"] : op[0] ? y["throw"] || ((t = y["return"]) && t.call(y), 0) : y.next) && !(t = t.call(y, op[1])).done) return t;
|
||||||
|
if (y = 0, t) op = [op[0] & 2, t.value];
|
||||||
|
switch (op[0]) {
|
||||||
|
case 0: case 1: t = op; break;
|
||||||
|
case 4: _.label++; return { value: op[1], done: false };
|
||||||
|
case 5: _.label++; y = op[1]; op = [0]; continue;
|
||||||
|
case 7: op = _.ops.pop(); _.trys.pop(); continue;
|
||||||
|
default:
|
||||||
|
if (!(t = _.trys, t = t.length > 0 && t[t.length - 1]) && (op[0] === 6 || op[0] === 2)) { _ = 0; continue; }
|
||||||
|
if (op[0] === 3 && (!t || (op[1] > t[0] && op[1] < t[3]))) { _.label = op[1]; break; }
|
||||||
|
if (op[0] === 6 && _.label < t[1]) { _.label = t[1]; t = op; break; }
|
||||||
|
if (t && _.label < t[2]) { _.label = t[2]; _.ops.push(op); break; }
|
||||||
|
if (t[2]) _.ops.pop();
|
||||||
|
_.trys.pop(); continue;
|
||||||
|
}
|
||||||
|
op = body.call(thisArg, _);
|
||||||
|
} catch (e) { op = [6, e]; y = 0; } finally { f = t = 0; }
|
||||||
|
if (op[0] & 5) throw op[1]; return { value: op[0] ? op[1] : void 0, done: true };
|
||||||
|
}
|
||||||
|
};
|
||||||
|
var __spreadArray = (this && this.__spreadArray) || function (to, from, pack) {
|
||||||
|
if (pack || arguments.length === 2) for (var i = 0, l = from.length, ar; i < l; i++) {
|
||||||
|
if (ar || !(i in from)) {
|
||||||
|
if (!ar) ar = Array.prototype.slice.call(from, 0, i);
|
||||||
|
ar[i] = from[i];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return to.concat(ar || Array.prototype.slice.call(from));
|
||||||
|
};
|
||||||
|
import Keyv from 'keyv';
|
||||||
|
import pTimeout from 'p-timeout';
|
||||||
|
import QuickLRU from 'quick-lru';
|
||||||
|
import { v4 as uuidv4 } from 'uuid';
|
||||||
|
import * as tokenizer from './tokenizer.js';
|
||||||
|
import * as types from './types.js';
|
||||||
|
import globalFetch from 'node-fetch';
|
||||||
|
var CHATGPT_MODEL = 'qwen-turbo'; // qwen-plus
|
||||||
|
var USER_LABEL_DEFAULT = 'User';
|
||||||
|
var ASSISTANT_LABEL_DEFAULT = '同义千问';
|
||||||
|
var QwenApi = /** @class */ (function () {
|
||||||
|
/**
|
||||||
|
* Creates a new client wrapper around Qwen's chat completion API, mimicing the official ChatGPT webapp's functionality as closely as possible.
|
||||||
|
*
|
||||||
|
* @param opts
|
||||||
|
*/
|
||||||
|
function QwenApi(opts) {
|
||||||
|
var apiKey = opts.apiKey, _a = opts.apiBaseUrl, apiBaseUrl = _a === void 0 ? 'https://dashscope.aliyuncs.com/api/v1' : _a, _b = opts.debug, debug = _b === void 0 ? false : _b, messageStore = opts.messageStore, completionParams = opts.completionParams, parameters = opts.parameters, systemMessage = opts.systemMessage, getMessageById = opts.getMessageById, upsertMessage = opts.upsertMessage, _c = opts.fetch, fetch = _c === void 0 ? globalFetch : _c;
|
||||||
|
this._apiKey = apiKey;
|
||||||
|
this._apiBaseUrl = apiBaseUrl;
|
||||||
|
this._debug = !!debug;
|
||||||
|
this._fetch = fetch;
|
||||||
|
this._completionParams = __assign({ model: CHATGPT_MODEL, parameters: __assign({ top_p: 0.5, top_k: 50, temperature: 1.0, seed: 114514, enable_search: true, result_format: "text", incremental_output: false }, parameters) }, completionParams);
|
||||||
|
this._systemMessage = systemMessage;
|
||||||
|
if (this._systemMessage === undefined) {
|
||||||
|
var currentDate = new Date().toISOString().split('T')[0];
|
||||||
|
this._systemMessage = "You are ChatGPT, a large language model trained by Qwen. Answer as concisely as possible.\nKnowledge cutoff: 2021-09-01\nCurrent date: ".concat(currentDate);
|
||||||
|
}
|
||||||
|
this._getMessageById = getMessageById !== null && getMessageById !== void 0 ? getMessageById : this._defaultGetMessageById;
|
||||||
|
this._upsertMessage = upsertMessage !== null && upsertMessage !== void 0 ? upsertMessage : this._defaultUpsertMessage;
|
||||||
|
if (messageStore) {
|
||||||
|
this._messageStore = messageStore;
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
this._messageStore = new Keyv({
|
||||||
|
store: new QuickLRU({ maxSize: 10000 })
|
||||||
|
});
|
||||||
|
}
|
||||||
|
if (!this._apiKey) {
|
||||||
|
throw new Error('Qwen missing required apiKey');
|
||||||
|
}
|
||||||
|
if (!this._fetch) {
|
||||||
|
throw new Error('Invalid environment; fetch is not defined');
|
||||||
|
}
|
||||||
|
if (typeof this._fetch !== 'function') {
|
||||||
|
throw new Error('Invalid "fetch" is not a function');
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Sends a message to the Qwen chat completions endpoint, waits for the response
|
||||||
|
* to resolve, and returns the response.
|
||||||
|
*
|
||||||
|
* If you want your response to have historical context, you must provide a valid `parentMessageId`.
|
||||||
|
*
|
||||||
|
* If you want to receive a stream of partial responses, use `opts.onProgress`.
|
||||||
|
*
|
||||||
|
* Set `debug: true` in the `ChatGPTAPI` constructor to log more info on the full prompt sent to the Qwen chat completions API. You can override the `systemMessage` in `opts` to customize the assistant's instructions.
|
||||||
|
*
|
||||||
|
* @param message - The prompt message to send
|
||||||
|
* @param opts.parentMessageId - Optional ID of the previous message in the conversation (defaults to `undefined`)
|
||||||
|
* @param opts.conversationId - Optional ID of the conversation (defaults to `undefined`)
|
||||||
|
* @param opts.messageId - Optional ID of the message to send (defaults to a random UUID)
|
||||||
|
* @param opts.systemMessage - Optional override for the chat "system message" which acts as instructions to the model (defaults to the ChatGPT system message)
|
||||||
|
* @param opts.timeoutMs - Optional timeout in milliseconds (defaults to no timeout)
|
||||||
|
* @param opts.onProgress - Optional callback which will be invoked every time the partial response is updated
|
||||||
|
* @param opts.abortSignal - Optional callback used to abort the underlying `fetch` call using an [AbortController](https://developer.mozilla.org/en-US/docs/Web/API/AbortController)
|
||||||
|
* @param completionParams - Optional overrides to send to the [Qwen chat completion API](https://platform.openai.com/docs/api-reference/chat/create). Options like `temperature` and `presence_penalty` can be tweaked to change the personality of the assistant.
|
||||||
|
*
|
||||||
|
* @returns The response from ChatGPT
|
||||||
|
*/
|
||||||
|
QwenApi.prototype.sendMessage = function (text, opts, role) {
|
||||||
|
if (opts === void 0) { opts = {}; }
|
||||||
|
if (role === void 0) { role = 'user'; }
|
||||||
|
return __awaiter(this, void 0, void 0, function () {
|
||||||
|
var parentMessageId, _a, messageId, timeoutMs, completionParams, conversationId, abortSignal, abortController, message, latestQuestion, _b, messages, maxTokens, numTokens, result, responseP;
|
||||||
|
var _this = this;
|
||||||
|
return __generator(this, function (_c) {
|
||||||
|
switch (_c.label) {
|
||||||
|
case 0:
|
||||||
|
parentMessageId = opts.parentMessageId, _a = opts.messageId, messageId = _a === void 0 ? uuidv4() : _a, timeoutMs = opts.timeoutMs, completionParams = opts.completionParams, conversationId = opts.conversationId;
|
||||||
|
abortSignal = opts.abortSignal;
|
||||||
|
abortController = null;
|
||||||
|
if (timeoutMs && !abortSignal) {
|
||||||
|
abortController = new AbortController();
|
||||||
|
abortSignal = abortController.signal;
|
||||||
|
}
|
||||||
|
message = {
|
||||||
|
role: role,
|
||||||
|
id: messageId,
|
||||||
|
conversationId: conversationId,
|
||||||
|
parentMessageId: parentMessageId,
|
||||||
|
text: text,
|
||||||
|
};
|
||||||
|
latestQuestion = message;
|
||||||
|
return [4 /*yield*/, this._buildMessages(text, role, opts, completionParams)];
|
||||||
|
case 1:
|
||||||
|
_b = _c.sent(), messages = _b.messages, maxTokens = _b.maxTokens, numTokens = _b.numTokens;
|
||||||
|
console.log("maxTokens: ".concat(maxTokens, ", numTokens: ").concat(numTokens));
|
||||||
|
result = {
|
||||||
|
role: 'assistant',
|
||||||
|
id: uuidv4(),
|
||||||
|
conversationId: conversationId,
|
||||||
|
parentMessageId: messageId,
|
||||||
|
text: undefined,
|
||||||
|
};
|
||||||
|
this._completionParams.input = { messages: messages };
|
||||||
|
responseP = new Promise(function (resolve, reject) { return __awaiter(_this, void 0, void 0, function () {
|
||||||
|
var url, headers, body, res, reason, msg, error, response, err_1;
|
||||||
|
return __generator(this, function (_a) {
|
||||||
|
switch (_a.label) {
|
||||||
|
case 0:
|
||||||
|
url = "".concat(this._apiBaseUrl, "/services/aigc/text-generation/generation");
|
||||||
|
headers = {
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
Authorization: "Bearer ".concat(this._apiKey)
|
||||||
|
};
|
||||||
|
body = __assign(__assign({}, this._completionParams), completionParams);
|
||||||
|
if (this._debug) {
|
||||||
|
console.log(JSON.stringify(body));
|
||||||
|
}
|
||||||
|
if (this._debug) {
|
||||||
|
console.log("sendMessage (".concat(numTokens, " tokens)"), body);
|
||||||
|
}
|
||||||
|
_a.label = 1;
|
||||||
|
case 1:
|
||||||
|
_a.trys.push([1, 6, , 7]);
|
||||||
|
return [4 /*yield*/, this._fetch(url, {
|
||||||
|
method: 'POST',
|
||||||
|
headers: headers,
|
||||||
|
body: JSON.stringify(body),
|
||||||
|
signal: abortSignal
|
||||||
|
})];
|
||||||
|
case 2:
|
||||||
|
res = _a.sent();
|
||||||
|
if (!!res.ok) return [3 /*break*/, 4];
|
||||||
|
return [4 /*yield*/, res.text()];
|
||||||
|
case 3:
|
||||||
|
reason = _a.sent();
|
||||||
|
msg = "Qwen error ".concat(res.status || res.statusText, ": ").concat(reason);
|
||||||
|
error = new types.ChatGPTError(msg, { cause: res });
|
||||||
|
error.statusCode = res.status;
|
||||||
|
error.statusText = res.statusText;
|
||||||
|
return [2 /*return*/, reject(error)];
|
||||||
|
case 4: return [4 /*yield*/, res.json()];
|
||||||
|
case 5:
|
||||||
|
response = _a.sent();
|
||||||
|
if (this._debug) {
|
||||||
|
console.log(response);
|
||||||
|
}
|
||||||
|
if (response === null || response === void 0 ? void 0 : response.request_id) {
|
||||||
|
result.id = response.request_id;
|
||||||
|
}
|
||||||
|
result.detail = response;
|
||||||
|
result.text = response.output.text;
|
||||||
|
return [2 /*return*/, resolve(result)];
|
||||||
|
case 6:
|
||||||
|
err_1 = _a.sent();
|
||||||
|
return [2 /*return*/, reject(err_1)];
|
||||||
|
case 7: return [2 /*return*/];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}); }).then(function (message) { return __awaiter(_this, void 0, void 0, function () {
|
||||||
|
return __generator(this, function (_a) {
|
||||||
|
return [2 /*return*/, Promise.all([
|
||||||
|
this._upsertMessage(latestQuestion),
|
||||||
|
this._upsertMessage(message)
|
||||||
|
]).then(function () { return message; })];
|
||||||
|
});
|
||||||
|
}); });
|
||||||
|
if (timeoutMs) {
|
||||||
|
if (abortController) {
|
||||||
|
// This will be called when a timeout occurs in order for us to forcibly
|
||||||
|
// ensure that the underlying HTTP request is aborted.
|
||||||
|
;
|
||||||
|
responseP.cancel = function () {
|
||||||
|
abortController.abort();
|
||||||
|
};
|
||||||
|
}
|
||||||
|
return [2 /*return*/, pTimeout(responseP, {
|
||||||
|
milliseconds: timeoutMs,
|
||||||
|
message: 'Qwen timed out waiting for response'
|
||||||
|
})];
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
return [2 /*return*/, responseP];
|
||||||
|
}
|
||||||
|
return [2 /*return*/];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
Object.defineProperty(QwenApi.prototype, "apiKey", {
|
||||||
|
get: function () {
|
||||||
|
return this._apiKey;
|
||||||
|
},
|
||||||
|
set: function (apiKey) {
|
||||||
|
this._apiKey = apiKey;
|
||||||
|
},
|
||||||
|
enumerable: false,
|
||||||
|
configurable: true
|
||||||
|
});
|
||||||
|
QwenApi.prototype._buildMessages = function (text, role, opts, completionParams) {
|
||||||
|
return __awaiter(this, void 0, void 0, function () {
|
||||||
|
var _a, systemMessage, parentMessageId, userLabel, assistantLabel, maxNumTokens, messages, systemMessageOffset, nextMessages, functionToken, numTokens, prompt_1, nextNumTokensEstimate, _i, nextMessages_1, m1, _b, isValidPrompt, parentMessage, parentMessageRole, maxTokens;
|
||||||
|
return __generator(this, function (_c) {
|
||||||
|
switch (_c.label) {
|
||||||
|
case 0:
|
||||||
|
_a = opts.systemMessage, systemMessage = _a === void 0 ? this._systemMessage : _a;
|
||||||
|
parentMessageId = opts.parentMessageId;
|
||||||
|
userLabel = USER_LABEL_DEFAULT;
|
||||||
|
assistantLabel = ASSISTANT_LABEL_DEFAULT;
|
||||||
|
maxNumTokens = 6000;
|
||||||
|
messages = [];
|
||||||
|
if (systemMessage) {
|
||||||
|
messages.push({
|
||||||
|
role: 'system',
|
||||||
|
content: systemMessage
|
||||||
|
});
|
||||||
|
}
|
||||||
|
systemMessageOffset = messages.length;
|
||||||
|
nextMessages = text
|
||||||
|
? messages.concat([
|
||||||
|
{
|
||||||
|
role: role,
|
||||||
|
content: text
|
||||||
|
}
|
||||||
|
])
|
||||||
|
: messages;
|
||||||
|
functionToken = 0;
|
||||||
|
numTokens = functionToken;
|
||||||
|
_c.label = 1;
|
||||||
|
case 1:
|
||||||
|
prompt_1 = nextMessages
|
||||||
|
.reduce(function (prompt, message) {
|
||||||
|
switch (message.role) {
|
||||||
|
case 'system':
|
||||||
|
return prompt.concat(["Instructions:\n".concat(message.content)]);
|
||||||
|
case 'user':
|
||||||
|
return prompt.concat(["".concat(userLabel, ":\n").concat(message.content)]);
|
||||||
|
default:
|
||||||
|
return message.content ? prompt.concat(["".concat(assistantLabel, ":\n").concat(message.content)]) : prompt;
|
||||||
|
}
|
||||||
|
}, [])
|
||||||
|
.join('\n\n');
|
||||||
|
return [4 /*yield*/, this._getTokenCount(prompt_1)];
|
||||||
|
case 2:
|
||||||
|
nextNumTokensEstimate = _c.sent();
|
||||||
|
_i = 0, nextMessages_1 = nextMessages;
|
||||||
|
_c.label = 3;
|
||||||
|
case 3:
|
||||||
|
if (!(_i < nextMessages_1.length)) return [3 /*break*/, 6];
|
||||||
|
m1 = nextMessages_1[_i];
|
||||||
|
_b = nextNumTokensEstimate;
|
||||||
|
return [4 /*yield*/, this._getTokenCount('')];
|
||||||
|
case 4:
|
||||||
|
nextNumTokensEstimate = _b + _c.sent();
|
||||||
|
_c.label = 5;
|
||||||
|
case 5:
|
||||||
|
_i++;
|
||||||
|
return [3 /*break*/, 3];
|
||||||
|
case 6:
|
||||||
|
isValidPrompt = nextNumTokensEstimate + functionToken <= maxNumTokens;
|
||||||
|
if (prompt_1 && !isValidPrompt) {
|
||||||
|
return [3 /*break*/, 9];
|
||||||
|
}
|
||||||
|
messages = nextMessages;
|
||||||
|
numTokens = nextNumTokensEstimate + functionToken;
|
||||||
|
if (!isValidPrompt) {
|
||||||
|
return [3 /*break*/, 9];
|
||||||
|
}
|
||||||
|
if (!parentMessageId) {
|
||||||
|
return [3 /*break*/, 9];
|
||||||
|
}
|
||||||
|
return [4 /*yield*/, this._getMessageById(parentMessageId)];
|
||||||
|
case 7:
|
||||||
|
parentMessage = _c.sent();
|
||||||
|
if (!parentMessage) {
|
||||||
|
return [3 /*break*/, 9];
|
||||||
|
}
|
||||||
|
parentMessageRole = parentMessage.role || 'user';
|
||||||
|
nextMessages = nextMessages.slice(0, systemMessageOffset).concat(__spreadArray([
|
||||||
|
{
|
||||||
|
role: parentMessageRole,
|
||||||
|
content: parentMessage.text
|
||||||
|
}
|
||||||
|
], nextMessages.slice(systemMessageOffset), true));
|
||||||
|
parentMessageId = parentMessage.parentMessageId;
|
||||||
|
_c.label = 8;
|
||||||
|
case 8:
|
||||||
|
if (true) return [3 /*break*/, 1];
|
||||||
|
_c.label = 9;
|
||||||
|
case 9:
|
||||||
|
maxTokens = Math.max(1, Math.min(this._maxModelTokens - numTokens, this._maxResponseTokens));
|
||||||
|
return [2 /*return*/, { messages: messages, maxTokens: maxTokens, numTokens: numTokens }];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
QwenApi.prototype._getTokenCount = function (text) {
|
||||||
|
return __awaiter(this, void 0, void 0, function () {
|
||||||
|
return __generator(this, function (_a) {
|
||||||
|
if (!text) {
|
||||||
|
return [2 /*return*/, 0];
|
||||||
|
}
|
||||||
|
// TODO: use a better fix in the tokenizer
|
||||||
|
text = text.replace(/<\|endoftext\|>/g, '');
|
||||||
|
return [2 /*return*/, tokenizer.encode(text).length];
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
QwenApi.prototype._defaultGetMessageById = function (id) {
|
||||||
|
return __awaiter(this, void 0, void 0, function () {
|
||||||
|
var res;
|
||||||
|
return __generator(this, function (_a) {
|
||||||
|
switch (_a.label) {
|
||||||
|
case 0: return [4 /*yield*/, this._messageStore.get(id)];
|
||||||
|
case 1:
|
||||||
|
res = _a.sent();
|
||||||
|
return [2 /*return*/, res];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
QwenApi.prototype._defaultUpsertMessage = function (message) {
|
||||||
|
return __awaiter(this, void 0, void 0, function () {
|
||||||
|
return __generator(this, function (_a) {
|
||||||
|
switch (_a.label) {
|
||||||
|
case 0: return [4 /*yield*/, this._messageStore.set(message.request_id, message)];
|
||||||
|
case 1:
|
||||||
|
_a.sent();
|
||||||
|
return [2 /*return*/];
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
return QwenApi;
|
||||||
|
}());
|
||||||
|
export { QwenApi };
|
||||||
382
utils/alibaba/qwen-api.ts
Normal file
382
utils/alibaba/qwen-api.ts
Normal file
|
|
@ -0,0 +1,382 @@
|
||||||
|
import Keyv from 'keyv'
|
||||||
|
import pTimeout from 'p-timeout'
|
||||||
|
import QuickLRU from 'quick-lru'
|
||||||
|
import { v4 as uuidv4 } from 'uuid'
|
||||||
|
|
||||||
|
import * as tokenizer from './tokenizer'
|
||||||
|
import * as types from './types'
|
||||||
|
import globalFetch from 'node-fetch'
|
||||||
|
import {qwen, Role} from "./types";
|
||||||
|
|
||||||
|
const CHATGPT_MODEL = 'qwen-turbo' // qwen-plus
|
||||||
|
|
||||||
|
const USER_LABEL_DEFAULT = 'User'
|
||||||
|
const ASSISTANT_LABEL_DEFAULT = '同义千问'
|
||||||
|
|
||||||
|
export class QwenApi {
|
||||||
|
protected _apiKey: string
|
||||||
|
protected _apiBaseUrl: string
|
||||||
|
protected _debug: boolean
|
||||||
|
|
||||||
|
protected _systemMessage: string
|
||||||
|
protected _completionParams: Omit<
|
||||||
|
types.qwen.CreateChatCompletionRequest,
|
||||||
|
'messages' | 'n'
|
||||||
|
>
|
||||||
|
protected _maxModelTokens: number
|
||||||
|
protected _maxResponseTokens: number
|
||||||
|
protected _fetch: types.FetchFn
|
||||||
|
|
||||||
|
protected _getMessageById: types.GetMessageByIdFunction
|
||||||
|
protected _upsertMessage: types.UpsertMessageFunction
|
||||||
|
|
||||||
|
protected _messageStore: Keyv<types.ChatMessage>
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a new client wrapper around Qwen's chat completion API, mimicing the official ChatGPT webapp's functionality as closely as possible.
|
||||||
|
*
|
||||||
|
* @param opts
|
||||||
|
*/
|
||||||
|
constructor(opts: types.QWenAPIOptions) {
|
||||||
|
const {
|
||||||
|
apiKey,
|
||||||
|
apiBaseUrl = 'https://dashscope.aliyuncs.com/api/v1',
|
||||||
|
debug = false,
|
||||||
|
messageStore,
|
||||||
|
completionParams,
|
||||||
|
parameters,
|
||||||
|
systemMessage,
|
||||||
|
getMessageById,
|
||||||
|
upsertMessage,
|
||||||
|
fetch = globalFetch
|
||||||
|
} = opts
|
||||||
|
|
||||||
|
this._apiKey = apiKey
|
||||||
|
this._apiBaseUrl = apiBaseUrl
|
||||||
|
this._debug = !!debug
|
||||||
|
this._fetch = fetch
|
||||||
|
|
||||||
|
this._completionParams = {
|
||||||
|
model: CHATGPT_MODEL,
|
||||||
|
parameters: {
|
||||||
|
top_p: 0.5,
|
||||||
|
top_k: 50,
|
||||||
|
temperature: 1.0,
|
||||||
|
seed: 114514,
|
||||||
|
enable_search: true,
|
||||||
|
result_format: "text",
|
||||||
|
incremental_output: false,
|
||||||
|
...parameters
|
||||||
|
},
|
||||||
|
...completionParams
|
||||||
|
}
|
||||||
|
|
||||||
|
this._systemMessage = systemMessage
|
||||||
|
|
||||||
|
if (this._systemMessage === undefined) {
|
||||||
|
const currentDate = new Date().toISOString().split('T')[0]
|
||||||
|
this._systemMessage = `You are ChatGPT, a large language model trained by Qwen. Answer as concisely as possible.\nKnowledge cutoff: 2021-09-01\nCurrent date: ${currentDate}`
|
||||||
|
}
|
||||||
|
|
||||||
|
this._getMessageById = getMessageById ?? this._defaultGetMessageById
|
||||||
|
this._upsertMessage = upsertMessage ?? this._defaultUpsertMessage
|
||||||
|
|
||||||
|
if (messageStore) {
|
||||||
|
this._messageStore = messageStore
|
||||||
|
} else {
|
||||||
|
this._messageStore = new Keyv<types.ChatMessage, any>({
|
||||||
|
store: new QuickLRU<string, types.ChatMessage>({ maxSize: 10000 })
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!this._apiKey) {
|
||||||
|
throw new Error('Qwen missing required apiKey')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!this._fetch) {
|
||||||
|
throw new Error('Invalid environment; fetch is not defined')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (typeof this._fetch !== 'function') {
|
||||||
|
throw new Error('Invalid "fetch" is not a function')
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sends a message to the Qwen chat completions endpoint, waits for the response
|
||||||
|
* to resolve, and returns the response.
|
||||||
|
*
|
||||||
|
* If you want your response to have historical context, you must provide a valid `parentMessageId`.
|
||||||
|
*
|
||||||
|
* If you want to receive a stream of partial responses, use `opts.onProgress`.
|
||||||
|
*
|
||||||
|
* Set `debug: true` in the `ChatGPTAPI` constructor to log more info on the full prompt sent to the Qwen chat completions API. You can override the `systemMessage` in `opts` to customize the assistant's instructions.
|
||||||
|
*
|
||||||
|
* @param message - The prompt message to send
|
||||||
|
* @param opts.parentMessageId - Optional ID of the previous message in the conversation (defaults to `undefined`)
|
||||||
|
* @param opts.conversationId - Optional ID of the conversation (defaults to `undefined`)
|
||||||
|
* @param opts.messageId - Optional ID of the message to send (defaults to a random UUID)
|
||||||
|
* @param opts.systemMessage - Optional override for the chat "system message" which acts as instructions to the model (defaults to the ChatGPT system message)
|
||||||
|
* @param opts.timeoutMs - Optional timeout in milliseconds (defaults to no timeout)
|
||||||
|
* @param opts.onProgress - Optional callback which will be invoked every time the partial response is updated
|
||||||
|
* @param opts.abortSignal - Optional callback used to abort the underlying `fetch` call using an [AbortController](https://developer.mozilla.org/en-US/docs/Web/API/AbortController)
|
||||||
|
* @param completionParams - Optional overrides to send to the [Qwen chat completion API](https://platform.openai.com/docs/api-reference/chat/create). Options like `temperature` and `presence_penalty` can be tweaked to change the personality of the assistant.
|
||||||
|
*
|
||||||
|
* @returns The response from ChatGPT
|
||||||
|
*/
|
||||||
|
async sendMessage(
|
||||||
|
text: string,
|
||||||
|
opts: types.SendMessageOptions = {},
|
||||||
|
role: Role = 'user',
|
||||||
|
): Promise<types.ChatMessage> {
|
||||||
|
const {
|
||||||
|
parentMessageId,
|
||||||
|
messageId = uuidv4(),
|
||||||
|
timeoutMs,
|
||||||
|
completionParams,
|
||||||
|
conversationId
|
||||||
|
} = opts
|
||||||
|
|
||||||
|
let { abortSignal } = opts
|
||||||
|
|
||||||
|
let abortController: AbortController = null
|
||||||
|
if (timeoutMs && !abortSignal) {
|
||||||
|
abortController = new AbortController()
|
||||||
|
abortSignal = abortController.signal
|
||||||
|
}
|
||||||
|
|
||||||
|
const message: types.ChatMessage = {
|
||||||
|
role,
|
||||||
|
id: messageId,
|
||||||
|
conversationId,
|
||||||
|
parentMessageId,
|
||||||
|
text,
|
||||||
|
}
|
||||||
|
|
||||||
|
const latestQuestion = message
|
||||||
|
|
||||||
|
const { messages, maxTokens, numTokens } = await this._buildMessages(
|
||||||
|
text,
|
||||||
|
role,
|
||||||
|
opts,
|
||||||
|
completionParams
|
||||||
|
)
|
||||||
|
console.log(`maxTokens: ${maxTokens}, numTokens: ${numTokens}`)
|
||||||
|
const result: types.ChatMessage = {
|
||||||
|
role: 'assistant',
|
||||||
|
id: uuidv4(),
|
||||||
|
conversationId,
|
||||||
|
parentMessageId: messageId,
|
||||||
|
text: undefined,
|
||||||
|
}
|
||||||
|
this._completionParams.input = { messages }
|
||||||
|
const responseP = new Promise<types.ChatMessage>(
|
||||||
|
async (resolve, reject) => {
|
||||||
|
const url = `${this._apiBaseUrl}/services/aigc/text-generation/generation`
|
||||||
|
const headers = {
|
||||||
|
'Content-Type': 'application/json',
|
||||||
|
Authorization: `Bearer ${this._apiKey}`
|
||||||
|
}
|
||||||
|
const body = {
|
||||||
|
...this._completionParams,
|
||||||
|
...completionParams
|
||||||
|
}
|
||||||
|
if (this._debug) {
|
||||||
|
console.log(JSON.stringify(body))
|
||||||
|
}
|
||||||
|
|
||||||
|
if (this._debug) {
|
||||||
|
console.log(`sendMessage (${numTokens} tokens)`, body)
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
const res = await this._fetch(url, {
|
||||||
|
method: 'POST',
|
||||||
|
headers,
|
||||||
|
body: JSON.stringify(body),
|
||||||
|
signal: abortSignal
|
||||||
|
})
|
||||||
|
|
||||||
|
if (!res.ok) {
|
||||||
|
const reason = await res.text()
|
||||||
|
const msg = `Qwen error ${
|
||||||
|
res.status || res.statusText
|
||||||
|
}: ${reason}`
|
||||||
|
const error = new types.ChatGPTError(msg, { cause: res })
|
||||||
|
error.statusCode = res.status
|
||||||
|
error.statusText = res.statusText
|
||||||
|
return reject(error)
|
||||||
|
}
|
||||||
|
|
||||||
|
const response: types.qwen.CreateChatCompletionResponse =
|
||||||
|
await res.json()
|
||||||
|
if (this._debug) {
|
||||||
|
console.log(response)
|
||||||
|
}
|
||||||
|
|
||||||
|
if (response?.request_id) {
|
||||||
|
result.id = response.request_id
|
||||||
|
}
|
||||||
|
result.detail = response
|
||||||
|
result.text = response.output.text
|
||||||
|
return resolve(result)
|
||||||
|
} catch (err) {
|
||||||
|
return reject(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
).then(async (message) => {
|
||||||
|
return Promise.all([
|
||||||
|
this._upsertMessage(latestQuestion),
|
||||||
|
this._upsertMessage(message)
|
||||||
|
]).then(() => message)
|
||||||
|
})
|
||||||
|
|
||||||
|
if (timeoutMs) {
|
||||||
|
if (abortController) {
|
||||||
|
// This will be called when a timeout occurs in order for us to forcibly
|
||||||
|
// ensure that the underlying HTTP request is aborted.
|
||||||
|
;(responseP as any).cancel = () => {
|
||||||
|
abortController.abort()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return pTimeout(responseP, {
|
||||||
|
milliseconds: timeoutMs,
|
||||||
|
message: 'Qwen timed out waiting for response'
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
return responseP
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
get apiKey(): string {
|
||||||
|
return this._apiKey
|
||||||
|
}
|
||||||
|
|
||||||
|
set apiKey(apiKey: string) {
|
||||||
|
this._apiKey = apiKey
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
protected async _buildMessages(text: string, role: Role, opts: types.SendMessageOptions, completionParams: Partial<
|
||||||
|
Omit<qwen.CreateChatCompletionRequest, 'messages' | 'n' | 'stream'>
|
||||||
|
>) {
|
||||||
|
const { systemMessage = this._systemMessage } = opts
|
||||||
|
let { parentMessageId } = opts
|
||||||
|
|
||||||
|
const userLabel = USER_LABEL_DEFAULT
|
||||||
|
const assistantLabel = ASSISTANT_LABEL_DEFAULT
|
||||||
|
|
||||||
|
// fix number of qwen
|
||||||
|
const maxNumTokens = 6000
|
||||||
|
let messages: types.qwen.ChatCompletionRequestMessage[] = []
|
||||||
|
|
||||||
|
if (systemMessage) {
|
||||||
|
messages.push({
|
||||||
|
role: 'system',
|
||||||
|
content: systemMessage
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
const systemMessageOffset = messages.length
|
||||||
|
let nextMessages = text
|
||||||
|
? messages.concat([
|
||||||
|
{
|
||||||
|
role,
|
||||||
|
content: text
|
||||||
|
}
|
||||||
|
])
|
||||||
|
: messages
|
||||||
|
|
||||||
|
let functionToken = 0
|
||||||
|
|
||||||
|
let numTokens = functionToken
|
||||||
|
|
||||||
|
do {
|
||||||
|
const prompt = nextMessages
|
||||||
|
.reduce((prompt, message) => {
|
||||||
|
switch (message.role) {
|
||||||
|
case 'system':
|
||||||
|
return prompt.concat([`Instructions:\n${message.content}`])
|
||||||
|
case 'user':
|
||||||
|
return prompt.concat([`${userLabel}:\n${message.content}`])
|
||||||
|
default:
|
||||||
|
return message.content ? prompt.concat([`${assistantLabel}:\n${message.content}`]) : prompt
|
||||||
|
}
|
||||||
|
}, [] as string[])
|
||||||
|
.join('\n\n')
|
||||||
|
|
||||||
|
let nextNumTokensEstimate = await this._getTokenCount(prompt)
|
||||||
|
|
||||||
|
for (const m1 of nextMessages) {
|
||||||
|
nextNumTokensEstimate += await this._getTokenCount('')
|
||||||
|
}
|
||||||
|
|
||||||
|
const isValidPrompt = nextNumTokensEstimate + functionToken <= maxNumTokens
|
||||||
|
|
||||||
|
if (prompt && !isValidPrompt) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
messages = nextMessages
|
||||||
|
numTokens = nextNumTokensEstimate + functionToken
|
||||||
|
|
||||||
|
if (!isValidPrompt) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!parentMessageId) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
const parentMessage = await this._getMessageById(parentMessageId)
|
||||||
|
if (!parentMessage) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
const parentMessageRole = parentMessage.role || 'user'
|
||||||
|
|
||||||
|
nextMessages = nextMessages.slice(0, systemMessageOffset).concat([
|
||||||
|
{
|
||||||
|
role: parentMessageRole,
|
||||||
|
content: parentMessage.text
|
||||||
|
},
|
||||||
|
...nextMessages.slice(systemMessageOffset)
|
||||||
|
])
|
||||||
|
|
||||||
|
parentMessageId = parentMessage.parentMessageId
|
||||||
|
|
||||||
|
} while (true)
|
||||||
|
|
||||||
|
// Use up to 4096 tokens (prompt + response), but try to leave 1000 tokens
|
||||||
|
// for the response.
|
||||||
|
const maxTokens = Math.max(
|
||||||
|
1,
|
||||||
|
Math.min(this._maxModelTokens - numTokens, this._maxResponseTokens)
|
||||||
|
)
|
||||||
|
|
||||||
|
return { messages, maxTokens, numTokens }
|
||||||
|
}
|
||||||
|
|
||||||
|
protected async _getTokenCount(text: string) {
|
||||||
|
if (!text) {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
// TODO: use a better fix in the tokenizer
|
||||||
|
text = text.replace(/<\|endoftext\|>/g, '')
|
||||||
|
|
||||||
|
return tokenizer.encode(text).length
|
||||||
|
}
|
||||||
|
|
||||||
|
protected async _defaultGetMessageById(
|
||||||
|
id: string
|
||||||
|
): Promise<types.ChatMessage> {
|
||||||
|
const res = await this._messageStore.get(id)
|
||||||
|
return res
|
||||||
|
}
|
||||||
|
|
||||||
|
protected async _defaultUpsertMessage(
|
||||||
|
message: types.ChatMessage
|
||||||
|
): Promise<void> {
|
||||||
|
await this._messageStore.set(message.request_id, message)
|
||||||
|
}
|
||||||
|
}
|
||||||
6
utils/alibaba/tokenizer.js
Normal file
6
utils/alibaba/tokenizer.js
Normal file
|
|
@ -0,0 +1,6 @@
|
||||||
|
import { getEncoding } from 'js-tiktoken';
|
||||||
|
// TODO: make this configurable
|
||||||
|
var tokenizer = getEncoding('cl100k_base');
|
||||||
|
export function encode(input) {
|
||||||
|
return new Uint32Array(tokenizer.encode(input));
|
||||||
|
}
|
||||||
8
utils/alibaba/tokenizer.ts
Normal file
8
utils/alibaba/tokenizer.ts
Normal file
|
|
@ -0,0 +1,8 @@
|
||||||
|
import { getEncoding } from 'js-tiktoken'
|
||||||
|
|
||||||
|
// TODO: make this configurable
|
||||||
|
const tokenizer = getEncoding('cl100k_base')
|
||||||
|
|
||||||
|
export function encode(input: string): Uint32Array {
|
||||||
|
return new Uint32Array(tokenizer.encode(input))
|
||||||
|
}
|
||||||
5
utils/alibaba/tsconfig.json
Normal file
5
utils/alibaba/tsconfig.json
Normal file
|
|
@ -0,0 +1,5 @@
|
||||||
|
{
|
||||||
|
"compilerOptions": {
|
||||||
|
"module": "es2020"
|
||||||
|
}
|
||||||
|
}
|
||||||
26
utils/alibaba/types.js
Normal file
26
utils/alibaba/types.js
Normal file
|
|
@ -0,0 +1,26 @@
|
||||||
|
var __extends = (this && this.__extends) || (function () {
|
||||||
|
var extendStatics = function (d, b) {
|
||||||
|
extendStatics = Object.setPrototypeOf ||
|
||||||
|
({ __proto__: [] } instanceof Array && function (d, b) { d.__proto__ = b; }) ||
|
||||||
|
function (d, b) { for (var p in b) if (Object.prototype.hasOwnProperty.call(b, p)) d[p] = b[p]; };
|
||||||
|
return extendStatics(d, b);
|
||||||
|
};
|
||||||
|
return function (d, b) {
|
||||||
|
if (typeof b !== "function" && b !== null)
|
||||||
|
throw new TypeError("Class extends value " + String(b) + " is not a constructor or null");
|
||||||
|
extendStatics(d, b);
|
||||||
|
function __() { this.constructor = d; }
|
||||||
|
d.prototype = b === null ? Object.create(b) : (__.prototype = b.prototype, new __());
|
||||||
|
};
|
||||||
|
})();
|
||||||
|
var ChatGPTError = /** @class */ (function (_super) {
|
||||||
|
__extends(ChatGPTError, _super);
|
||||||
|
function ChatGPTError() {
|
||||||
|
return _super !== null && _super.apply(this, arguments) || this;
|
||||||
|
}
|
||||||
|
return ChatGPTError;
|
||||||
|
}(Error));
|
||||||
|
export { ChatGPTError };
|
||||||
|
export var qwen;
|
||||||
|
(function (qwen) {
|
||||||
|
})(qwen || (qwen = {}));
|
||||||
313
utils/alibaba/types.ts
Normal file
313
utils/alibaba/types.ts
Normal file
|
|
@ -0,0 +1,313 @@
|
||||||
|
import Keyv from 'keyv'
|
||||||
|
|
||||||
|
export type Role = 'user' | 'assistant' | 'system'
|
||||||
|
|
||||||
|
export type FetchFn = typeof fetch
|
||||||
|
|
||||||
|
export type QWenAPIOptions = {
|
||||||
|
apiKey: string
|
||||||
|
|
||||||
|
/** @defaultValue `'https://dashscope.aliyuncs.com/api/v1'` **/
|
||||||
|
apiBaseUrl?: string
|
||||||
|
|
||||||
|
apiOrg?: string
|
||||||
|
|
||||||
|
/** @defaultValue `false` **/
|
||||||
|
debug?: boolean
|
||||||
|
|
||||||
|
completionParams?: Partial<
|
||||||
|
Omit<qwen.CreateChatCompletionRequest, 'messages' | 'n' | 'stream'>
|
||||||
|
>
|
||||||
|
parameters?: qwen.QWenParameters,
|
||||||
|
|
||||||
|
systemMessage?: string
|
||||||
|
|
||||||
|
messageStore?: Keyv
|
||||||
|
getMessageById?: GetMessageByIdFunction
|
||||||
|
upsertMessage?: UpsertMessageFunction
|
||||||
|
|
||||||
|
fetch?: FetchFn
|
||||||
|
}
|
||||||
|
|
||||||
|
export type SendMessageOptions = {
|
||||||
|
/**
|
||||||
|
* function role name
|
||||||
|
*/
|
||||||
|
name?: string
|
||||||
|
messageId?: string
|
||||||
|
stream?: boolean
|
||||||
|
systemMessage?: string
|
||||||
|
parentMessageId?: string
|
||||||
|
conversationId?: string
|
||||||
|
timeoutMs?: number
|
||||||
|
onProgress?: (partialResponse: ChatMessage) => void
|
||||||
|
abortSignal?: AbortSignal
|
||||||
|
completionParams?: Partial<
|
||||||
|
Omit<qwen.CreateChatCompletionRequest, 'messages' | 'n' | 'stream'>
|
||||||
|
>
|
||||||
|
}
|
||||||
|
|
||||||
|
export type MessageActionType = 'next' | 'variant'
|
||||||
|
|
||||||
|
export type SendMessageBrowserOptions = {
|
||||||
|
conversationId?: string
|
||||||
|
parentMessageId?: string
|
||||||
|
messageId?: string
|
||||||
|
action?: MessageActionType
|
||||||
|
timeoutMs?: number
|
||||||
|
onProgress?: (partialResponse: ChatMessage) => void
|
||||||
|
abortSignal?: AbortSignal
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ChatMessage {
|
||||||
|
id: string
|
||||||
|
text: string
|
||||||
|
role: Role
|
||||||
|
parentMessageId?: string
|
||||||
|
conversationId?: string
|
||||||
|
detail?:
|
||||||
|
| qwen.CreateChatCompletionResponse
|
||||||
|
| CreateChatCompletionStreamResponse
|
||||||
|
}
|
||||||
|
|
||||||
|
export class ChatGPTError extends Error {
|
||||||
|
statusCode?: number
|
||||||
|
statusText?: string
|
||||||
|
isFinal?: boolean
|
||||||
|
accountId?: string
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Returns a chat message from a store by it's ID (or null if not found). */
|
||||||
|
export type GetMessageByIdFunction = (id: string) => Promise<ChatMessage>
|
||||||
|
|
||||||
|
/** Upserts a chat message to a store. */
|
||||||
|
export type UpsertMessageFunction = (message: ChatMessage) => Promise<void>
|
||||||
|
|
||||||
|
export interface CreateChatCompletionStreamResponse
|
||||||
|
extends openai.CreateChatCompletionDeltaResponse {
|
||||||
|
usage: CreateCompletionStreamResponseUsage
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface CreateCompletionStreamResponseUsage
|
||||||
|
extends openai.CreateCompletionResponseUsage {
|
||||||
|
estimated: true
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* https://chat.openapi.com/backend-api/conversation
|
||||||
|
*/
|
||||||
|
export type ConversationJSONBody = {
|
||||||
|
/**
|
||||||
|
* The action to take
|
||||||
|
*/
|
||||||
|
action: string
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The ID of the conversation
|
||||||
|
*/
|
||||||
|
conversation_id?: string
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Prompts to provide
|
||||||
|
*/
|
||||||
|
messages: Prompt[]
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The model to use
|
||||||
|
*/
|
||||||
|
model: string
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The parent message ID
|
||||||
|
*/
|
||||||
|
parent_message_id: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export type Prompt = {
|
||||||
|
/**
|
||||||
|
* The content of the prompt
|
||||||
|
*/
|
||||||
|
content: PromptContent
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The ID of the prompt
|
||||||
|
*/
|
||||||
|
id: string
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The role played in the prompt
|
||||||
|
*/
|
||||||
|
role: Role
|
||||||
|
}
|
||||||
|
|
||||||
|
export type ContentType = 'text'
|
||||||
|
|
||||||
|
export type PromptContent = {
|
||||||
|
/**
|
||||||
|
* The content type of the prompt
|
||||||
|
*/
|
||||||
|
content_type: ContentType
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The parts to the prompt
|
||||||
|
*/
|
||||||
|
parts: string[]
|
||||||
|
}
|
||||||
|
|
||||||
|
export type ConversationResponseEvent = {
|
||||||
|
message?: Message
|
||||||
|
conversation_id?: string
|
||||||
|
error?: string | null
|
||||||
|
}
|
||||||
|
|
||||||
|
export type Message = {
|
||||||
|
id: string
|
||||||
|
content: MessageContent
|
||||||
|
role: Role
|
||||||
|
user: string | null
|
||||||
|
create_time: string | null
|
||||||
|
update_time: string | null
|
||||||
|
end_turn: null
|
||||||
|
weight: number
|
||||||
|
recipient: string
|
||||||
|
metadata: MessageMetadata
|
||||||
|
}
|
||||||
|
|
||||||
|
export type MessageContent = {
|
||||||
|
content_type: string
|
||||||
|
parts: string[]
|
||||||
|
}
|
||||||
|
|
||||||
|
export type MessageMetadata = any
|
||||||
|
|
||||||
|
export namespace qwen {
|
||||||
|
export interface CreateChatCompletionDeltaResponse {
|
||||||
|
id: string
|
||||||
|
object: 'chat.completion.chunk'
|
||||||
|
created: number
|
||||||
|
model: string
|
||||||
|
choices: [
|
||||||
|
{
|
||||||
|
delta: {
|
||||||
|
role: Role
|
||||||
|
content?: string,
|
||||||
|
function_call?: {name: string, arguments: string}
|
||||||
|
}
|
||||||
|
index: number
|
||||||
|
finish_reason: string | null
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @export
|
||||||
|
* @interface ChatCompletionRequestMessage
|
||||||
|
*/
|
||||||
|
export interface ChatCompletionRequestMessage {
|
||||||
|
/**
|
||||||
|
* The role of the author of this message.
|
||||||
|
* @type {string}
|
||||||
|
* @memberof ChatCompletionRequestMessage
|
||||||
|
*/
|
||||||
|
role: ChatCompletionRequestMessageRoleEnum
|
||||||
|
/**
|
||||||
|
* The contents of the message
|
||||||
|
* @type {string}
|
||||||
|
* @memberof ChatCompletionRequestMessage
|
||||||
|
*/
|
||||||
|
content: string
|
||||||
|
}
|
||||||
|
|
||||||
|
export declare const ChatCompletionRequestMessageRoleEnum: {
|
||||||
|
readonly System: 'system'
|
||||||
|
readonly User: 'user'
|
||||||
|
readonly Assistant: 'assistant'
|
||||||
|
}
|
||||||
|
export declare type ChatCompletionRequestMessageRoleEnum =
|
||||||
|
(typeof ChatCompletionRequestMessageRoleEnum)[keyof typeof ChatCompletionRequestMessageRoleEnum]
|
||||||
|
|
||||||
|
|
||||||
|
export interface QWenInput {
|
||||||
|
messages: Array<ChatCompletionRequestMessage>
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface QWenParameters {
|
||||||
|
result_format: string
|
||||||
|
top_p: number
|
||||||
|
top_k: number
|
||||||
|
seed: number
|
||||||
|
temperature: number
|
||||||
|
enable_search: boolean
|
||||||
|
incremental_output: boolean
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @export
|
||||||
|
* @interface CreateChatCompletionRequest
|
||||||
|
*/
|
||||||
|
export interface CreateChatCompletionRequest {
|
||||||
|
/**
|
||||||
|
* ID of the model to use. Currently, only `gpt-3.5-turbo` and `gpt-3.5-turbo-0301` are supported.
|
||||||
|
* @type {string}
|
||||||
|
* @memberof CreateChatCompletionRequest
|
||||||
|
*/
|
||||||
|
model: string
|
||||||
|
/**
|
||||||
|
* The messages to generate chat completions for, in the [chat format](/docs/guides/chat/introduction).
|
||||||
|
* @type {Array<ChatCompletionRequestMessage>}
|
||||||
|
* @memberof CreateChatCompletionRequest
|
||||||
|
*/
|
||||||
|
input?: QWenInput
|
||||||
|
|
||||||
|
parameters: QWenParameters
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @export
|
||||||
|
* @interface CreateChatCompletionResponse
|
||||||
|
*/
|
||||||
|
export interface CreateChatCompletionResponse {
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @type {string}
|
||||||
|
* @memberof CreateChatCompletionResponse
|
||||||
|
*/
|
||||||
|
request_id: string
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @type {QWenOutput}
|
||||||
|
* @memberof CreateChatCompletionResponse
|
||||||
|
*/
|
||||||
|
output: QWenOutput
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @type {CreateCompletionResponseUsage}
|
||||||
|
* @memberof CreateChatCompletionResponse
|
||||||
|
*/
|
||||||
|
usage?: CreateCompletionResponseUsage
|
||||||
|
}
|
||||||
|
export interface QWenOutput {
|
||||||
|
finish_reason: string
|
||||||
|
text: string
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @export
|
||||||
|
* @interface CreateCompletionResponseUsage
|
||||||
|
*/
|
||||||
|
export interface CreateCompletionResponseUsage {
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @type {number}
|
||||||
|
* @memberof CreateCompletionResponseUsage
|
||||||
|
*/
|
||||||
|
input_tokens: number
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @type {number}
|
||||||
|
* @memberof CreateCompletionResponseUsage
|
||||||
|
*/
|
||||||
|
output_tokens: number
|
||||||
|
}
|
||||||
|
}
|
||||||
71
utils/bilibili/wbi.js
Normal file
71
utils/bilibili/wbi.js
Normal file
|
|
@ -0,0 +1,71 @@
|
||||||
|
import md5 from 'md5'
|
||||||
|
import fetch from 'node-fetch'
|
||||||
|
|
||||||
|
const mixinKeyEncTab = [
|
||||||
|
46, 47, 18, 2, 53, 8, 23, 32, 15, 50, 10, 31, 58, 3, 45, 35, 27, 43, 5, 49,
|
||||||
|
33, 9, 42, 19, 29, 28, 14, 39, 12, 38, 41, 13, 37, 48, 7, 16, 24, 55, 40,
|
||||||
|
61, 26, 17, 0, 1, 60, 51, 30, 4, 22, 25, 54, 21, 56, 59, 6, 63, 57, 62, 11,
|
||||||
|
36, 20, 34, 44, 52
|
||||||
|
]
|
||||||
|
|
||||||
|
// 对 imgKey 和 subKey 进行字符顺序打乱编码
|
||||||
|
function getMixinKey (orig) {
|
||||||
|
let temp = ''
|
||||||
|
mixinKeyEncTab.forEach((n) => {
|
||||||
|
temp += orig[n]
|
||||||
|
})
|
||||||
|
return temp.slice(0, 32)
|
||||||
|
}
|
||||||
|
|
||||||
|
// 为请求参数进行 wbi 签名
|
||||||
|
function encWbi (params, imgKey, subKey) {
|
||||||
|
const mixinKey = getMixinKey(imgKey + subKey)
|
||||||
|
const currTime = Math.round(Date.now() / 1000)
|
||||||
|
const chrFilter = /[!'()*]/g
|
||||||
|
let query = []
|
||||||
|
Object.assign(params, { wts: currTime }) // 添加 wts 字段
|
||||||
|
// 按照 key 重排参数
|
||||||
|
Object.keys(params).sort().forEach((key) => {
|
||||||
|
query.push(
|
||||||
|
`${encodeURIComponent(key)}=${encodeURIComponent(
|
||||||
|
// 过滤 value 中的 "!'()*" 字符
|
||||||
|
params[key].toString().replace(chrFilter, '')
|
||||||
|
)}`
|
||||||
|
)
|
||||||
|
})
|
||||||
|
query = query.join('&')
|
||||||
|
const wbiSign = md5(query + mixinKey) // 计算 w_rid
|
||||||
|
return query + '&w_rid=' + wbiSign
|
||||||
|
}
|
||||||
|
|
||||||
|
// 获取最新的 img_key 和 sub_key
|
||||||
|
async function getWbiKeys () {
|
||||||
|
const resp = await fetch('https://api.bilibili.com/x/web-interface/nav')
|
||||||
|
const jsonContent = resp.data
|
||||||
|
const imgUrl = jsonContent.data.wbi_img.img_url
|
||||||
|
const subUrl = jsonContent.data.wbi_img.sub_url
|
||||||
|
|
||||||
|
return {
|
||||||
|
img_key: imgUrl.slice(
|
||||||
|
imgUrl.lastIndexOf('/') + 1,
|
||||||
|
imgUrl.lastIndexOf('.')
|
||||||
|
),
|
||||||
|
sub_key: subUrl.slice(
|
||||||
|
subUrl.lastIndexOf('/') + 1,
|
||||||
|
subUrl.lastIndexOf('.')
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getWbiKeys().then((wbi_keys) => {
|
||||||
|
// const query = encWbi(
|
||||||
|
// {
|
||||||
|
// foo: '114',
|
||||||
|
// bar: '514',
|
||||||
|
// baz: 1919810
|
||||||
|
// },
|
||||||
|
// wbi_keys.img_key,
|
||||||
|
// wbi_keys.sub_key
|
||||||
|
// )
|
||||||
|
// console.log(query)
|
||||||
|
// })
|
||||||
|
|
@ -186,6 +186,7 @@ export class ClaudeAIClient {
|
||||||
} else if (streamDataRes.status === 408) {
|
} else if (streamDataRes.status === 408) {
|
||||||
throw new Error('claude.ai响应超时,可能是回复文本太多,请调高超时时间重试')
|
throw new Error('claude.ai响应超时,可能是回复文本太多,请调高超时时间重试')
|
||||||
} else {
|
} else {
|
||||||
|
logger.error(streamDataRes.status, streamDataRes.body)
|
||||||
throw new Error('unknown error')
|
throw new Error('unknown error')
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
249
utils/common.js
249
utils/common.js
|
|
@ -1,5 +1,3 @@
|
||||||
// import { remark } from 'remark'
|
|
||||||
// import stripMarkdown from 'strip-markdown'
|
|
||||||
import { exec } from 'child_process'
|
import { exec } from 'child_process'
|
||||||
import lodash from 'lodash'
|
import lodash from 'lodash'
|
||||||
import fs from 'node:fs'
|
import fs from 'node:fs'
|
||||||
|
|
@ -15,12 +13,26 @@ import AzureTTS, { supportConfigurations as azureRoleList } from './tts/microsof
|
||||||
import { translate } from './translate.js'
|
import { translate } from './translate.js'
|
||||||
import uploadRecord from './uploadRecord.js'
|
import uploadRecord from './uploadRecord.js'
|
||||||
import Version from './version.js'
|
import Version from './version.js'
|
||||||
// export function markdownToText (markdown) {
|
import fetch from 'node-fetch'
|
||||||
// return remark()
|
let pdfjsLib
|
||||||
// .use(stripMarkdown)
|
try {
|
||||||
// .processSync(markdown ?? '')
|
pdfjsLib = (await import('pdfjs-dist')).default
|
||||||
// .toString()
|
} catch (err) {}
|
||||||
// }
|
|
||||||
|
let mammoth
|
||||||
|
try {
|
||||||
|
mammoth = (await import('mammoth')).default
|
||||||
|
} catch (err) {}
|
||||||
|
|
||||||
|
let XLSX
|
||||||
|
try {
|
||||||
|
XLSX = (await import('xlsx')).default
|
||||||
|
} catch (err) {}
|
||||||
|
|
||||||
|
let PPTX
|
||||||
|
try {
|
||||||
|
PPTX = (await import('nodejs-pptx')).default
|
||||||
|
} catch (err) {}
|
||||||
|
|
||||||
let _puppeteer
|
let _puppeteer
|
||||||
try {
|
try {
|
||||||
|
|
@ -61,12 +73,18 @@ export function randomString (length = 5) {
|
||||||
return str.substr(0, length)
|
return str.substr(0, length)
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function upsertMessage (message) {
|
export async function upsertMessage (message, suffix = '') {
|
||||||
await redis.set(`CHATGPT:MESSAGE:${message.id}`, JSON.stringify(message))
|
if (suffix) {
|
||||||
|
suffix = '_' + suffix
|
||||||
|
}
|
||||||
|
await redis.set(`CHATGPT:MESSAGE${suffix}:${message.id}`, JSON.stringify(message))
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function getMessageById (id) {
|
export async function getMessageById (id, suffix = '') {
|
||||||
let messageStr = await redis.get(`CHATGPT:MESSAGE:${id}`)
|
if (suffix) {
|
||||||
|
suffix = '_' + suffix
|
||||||
|
}
|
||||||
|
let messageStr = await redis.get(`CHATGPT:MESSAGE${suffix}:${id}`)
|
||||||
return JSON.parse(messageStr)
|
return JSON.parse(messageStr)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -833,7 +851,7 @@ export function getUin (e) {
|
||||||
else {
|
else {
|
||||||
Bot.uin.forEach((u) => {
|
Bot.uin.forEach((u) => {
|
||||||
if (Bot[u].self_id) {
|
if (Bot[u].self_id) {
|
||||||
return Bot[u].self_id
|
return Bot[u].self_id
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
return Bot.uin[Bot.uin.length - 1]
|
return Bot.uin[Bot.uin.length - 1]
|
||||||
|
|
@ -972,3 +990,208 @@ export function getUserSpeaker (userSetting) {
|
||||||
return userSetting.ttsRoleVoiceVox || Config.voicevoxTTSSpeaker
|
return userSetting.ttsRoleVoiceVox || Config.voicevoxTTSSpeaker
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @param url 要下载的文件链接
|
||||||
|
* @param destPath 目标路径,如received/abc.pdf. 目前如果文件名重复会覆盖。
|
||||||
|
* @param absolute 是否是绝对路径,默认为false,此时拼接在data/chatgpt下
|
||||||
|
* @returns {Promise<string>} 最终下载文件的存储位置
|
||||||
|
*/
|
||||||
|
export async function downloadFile (url, destPath, absolute = false) {
|
||||||
|
let response = await fetch(url)
|
||||||
|
if (!response.ok) {
|
||||||
|
throw new Error(`download file http error: status: ${response.status}`)
|
||||||
|
}
|
||||||
|
let dest = destPath
|
||||||
|
if (!absolute) {
|
||||||
|
const _path = process.cwd()
|
||||||
|
dest = path.join(_path, 'data', 'chatgpt', dest)
|
||||||
|
const lastLevelDirPath = path.dirname(dest)
|
||||||
|
mkdirs(lastLevelDirPath)
|
||||||
|
}
|
||||||
|
const fileStream = fs.createWriteStream(dest)
|
||||||
|
await new Promise((resolve, reject) => {
|
||||||
|
response.body.pipe(fileStream)
|
||||||
|
response.body.on('error', err => {
|
||||||
|
reject(err)
|
||||||
|
})
|
||||||
|
fileStream.on('finish', function () {
|
||||||
|
resolve()
|
||||||
|
})
|
||||||
|
})
|
||||||
|
logger.info(`File downloaded successfully! URL: ${url}, Destination: ${dest}`)
|
||||||
|
return dest
|
||||||
|
}
|
||||||
|
|
||||||
|
export function isPureText (filename) {
|
||||||
|
const ext = path.extname(filename).toLowerCase()
|
||||||
|
|
||||||
|
// List of file extensions that can be treated as pure text
|
||||||
|
const textFileExtensions = ['.txt', '.log', '.md', '.csv', '.html', '.css', '.js', '.json', '.xml', '.py', '.java', '.cpp', '.c', '.rb', '.php', '.sql', '.sh', '.pl', '.r', '.swift', '.go', '.ts', '.htm', '.yaml', '.yml', '.ini', '.properties', '.tsv']
|
||||||
|
|
||||||
|
// File types that require additional processing
|
||||||
|
const processingExtensions = ['.docx', '.pptx', '.xlsx', '.pdf', '.epub']
|
||||||
|
|
||||||
|
if (textFileExtensions.includes(ext)) {
|
||||||
|
return 'text'
|
||||||
|
} else if (processingExtensions.includes(ext)) {
|
||||||
|
// Return the file extension if additional processing is needed
|
||||||
|
return ext.replace('.', '')
|
||||||
|
} else {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 从文件中提取文本内容
|
||||||
|
* @param fileMsgElem MessageElem
|
||||||
|
* @returns {Promise<{}>} 提取的文本内容和文件名
|
||||||
|
*/
|
||||||
|
export async function extractContentFromFile (fileMsgElem, e) {
|
||||||
|
logger.info('filename: ' + fileMsgElem.name)
|
||||||
|
let fileType = isPureText(fileMsgElem.name)
|
||||||
|
if (fileType) {
|
||||||
|
// 可读的文件类型
|
||||||
|
let fileUrl = e.isGroup ? await e.group.getFileUrl(fileMsgElem.fid) : await e.friend.getFileUrl(fileMsgElem.fid)
|
||||||
|
let filePath = await downloadFile(fileUrl, path.join('received', fileMsgElem.name))
|
||||||
|
switch (fileType) {
|
||||||
|
case 'pdf': {
|
||||||
|
if (!pdfjsLib) {
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
const data = new Uint8Array(fs.readFileSync(filePath))
|
||||||
|
let loadingTask = pdfjsLib.getDocument(data)
|
||||||
|
try {
|
||||||
|
const pdfDocument = await loadingTask.promise
|
||||||
|
const numPages = pdfDocument.numPages
|
||||||
|
let pdfText = ''
|
||||||
|
|
||||||
|
// limit pages to prevent OOM or LLM down
|
||||||
|
let maxPage = 100
|
||||||
|
// Iterate through each page and extract text
|
||||||
|
for (let pageNum = 1; pageNum <= Math.min(numPages, maxPage); ++pageNum) {
|
||||||
|
const page = await pdfDocument.getPage(pageNum)
|
||||||
|
const textContent = await page.getTextContent()
|
||||||
|
const pageText = textContent.items.map(item => item.str).join(' ')
|
||||||
|
pdfText += pageText
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
content: pdfText,
|
||||||
|
name: fileMsgElem.name
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error reading PDF file:', error)
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case 'doc': {
|
||||||
|
logger.error('not supported file type now')
|
||||||
|
return ''
|
||||||
|
}
|
||||||
|
case 'docx': {
|
||||||
|
if (!mammoth) {
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
const { value } = await mammoth.extractRawText({ path: filePath })
|
||||||
|
return {
|
||||||
|
content: value,
|
||||||
|
name: fileMsgElem.name
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
logger.error('Error reading .docx file:', error)
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case 'xls': {
|
||||||
|
logger.error('not supported file type now')
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
case 'xlsx': {
|
||||||
|
if (!XLSX) {
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
const workbook = XLSX.readFile(filePath)
|
||||||
|
const sheetName = workbook.SheetNames[0] // Assuming the first sheet is the one you want to read
|
||||||
|
const sheet = workbook.Sheets[sheetName]
|
||||||
|
const data = XLSX.utils.sheet_to_json(sheet, { header: 1 })
|
||||||
|
|
||||||
|
// Convert the 2D array to plain text
|
||||||
|
return {
|
||||||
|
content: data.map(row => row.join('\t')).join('\n'),
|
||||||
|
name: fileMsgElem.name
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error reading .xlsx file:', error)
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case 'ppt': {
|
||||||
|
logger.error('not supported file type now')
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
case 'pptx': {
|
||||||
|
if (!PPTX) {
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
try {
|
||||||
|
let pptx = new PPTX.Composer()
|
||||||
|
await pptx.load(filePath)
|
||||||
|
let presentationContent = []
|
||||||
|
let slideNumber = 1
|
||||||
|
let maxSlideNumber = 60
|
||||||
|
while (slideNumber <= maxSlideNumber) {
|
||||||
|
let slide
|
||||||
|
try {
|
||||||
|
slide = pptx.getSlide(slideNumber)
|
||||||
|
} catch (error) {
|
||||||
|
// Slide number out of range, break the loop
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
let slideContent = []
|
||||||
|
|
||||||
|
// Iterate through slide elements and extract text content
|
||||||
|
slide.elements.forEach(element => {
|
||||||
|
if (element.text) {
|
||||||
|
slideContent.push(element.text)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
// Add slide content to the presentation content array
|
||||||
|
presentationContent.push(slideContent.join('\n'))
|
||||||
|
|
||||||
|
// Move to the next slide
|
||||||
|
slideNumber++
|
||||||
|
}
|
||||||
|
return {
|
||||||
|
content: presentationContent.join('\n'),
|
||||||
|
name: fileMsgElem.name
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error('Error reading .pptx file:', error)
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case 'epub': {
|
||||||
|
logger.error('not supported file type now')
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
default: {
|
||||||
|
// text type
|
||||||
|
const data = fs.readFileSync(filePath)
|
||||||
|
let text = String(data)
|
||||||
|
if (text) {
|
||||||
|
return {
|
||||||
|
content: text,
|
||||||
|
name: fileMsgElem.name
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -154,7 +154,15 @@ const defaultConfig = {
|
||||||
claudeAIUA: 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36',
|
claudeAIUA: 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36',
|
||||||
// trss配置
|
// trss配置
|
||||||
trssBotUin: '',
|
trssBotUin: '',
|
||||||
version: 'v2.7.6'
|
// 同义千问
|
||||||
|
qwenApiKey: '',
|
||||||
|
qwenModel: 'qwen-turbo',
|
||||||
|
qwenTopP: 0.5,
|
||||||
|
qwenTopK: 50,
|
||||||
|
qwenSeed: 0,
|
||||||
|
qwenTemperature: 1,
|
||||||
|
qwenEnableSearch: true,
|
||||||
|
version: 'v2.7.7'
|
||||||
}
|
}
|
||||||
const _path = process.cwd()
|
const _path = process.cwd()
|
||||||
let config = {}
|
let config = {}
|
||||||
|
|
|
||||||
|
|
@ -139,7 +139,7 @@ async function uploadRecord (recordUrl, ttsMode = 'vits-uma-genshin-honkai') {
|
||||||
2: 3,
|
2: 3,
|
||||||
5: {
|
5: {
|
||||||
1: Contactable.target,
|
1: Contactable.target,
|
||||||
2: getUin(e),
|
2: getUin(),
|
||||||
3: 0,
|
3: 0,
|
||||||
4: hash,
|
4: hash,
|
||||||
5: buf.length,
|
5: buf.length,
|
||||||
|
|
@ -183,7 +183,7 @@ async function uploadRecord (recordUrl, ttsMode = 'vits-uma-genshin-honkai') {
|
||||||
const fid = rsp[11].toBuffer()
|
const fid = rsp[11].toBuffer()
|
||||||
const b = core.pb.encode({
|
const b = core.pb.encode({
|
||||||
1: 4,
|
1: 4,
|
||||||
2: Bot.uin,
|
2: getUin(),
|
||||||
3: fid,
|
3: fid,
|
||||||
4: hash,
|
4: hash,
|
||||||
5: hash.toString('hex') + '.amr',
|
5: hash.toString('hex') + '.amr',
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue