Update API keys in config.json
This commit is contained in:
parent
77e36eee64
commit
96603eda55
1 changed files with 129 additions and 99 deletions
|
@ -253,33 +253,39 @@ module.exports = async (client, message) => {
|
||||||
|
|
||||||
// ChatBot
|
// ChatBot
|
||||||
if (chatbot.State && chatbot.AllowedChannels.includes(message.channel.name) && !chatbot.IgnoredUsers.includes(message.author.id)) {
|
if (chatbot.State && chatbot.AllowedChannels.includes(message.channel.name) && !chatbot.IgnoredUsers.includes(message.author.id)) {
|
||||||
|
// Check if the message is a direct message
|
||||||
|
if (message.channel.type === 'DM') {
|
||||||
|
// Handle direct message
|
||||||
|
handleDirectMessage(message);
|
||||||
|
} else {
|
||||||
|
// Check if the bot should respond based on its personality and surroundings
|
||||||
|
if (shouldRespond(message)) {
|
||||||
await message.channel.sendTyping();
|
await message.channel.sendTyping();
|
||||||
|
|
||||||
const openai = new openAI.OpenAI({ apiKey: config.OpenAIapiKey });
|
const openai = new openAI.OpenAI({ apiKey: config.OpenAIapiKey });
|
||||||
|
|
||||||
const question = message.content;
|
const question = message.content;
|
||||||
|
|
||||||
const completionPrompt = fs.readFileSync("./utils/prompts/completion.txt", "utf-8");
|
const completionPrompt = `
|
||||||
const prompt = completionPrompt.replaceAll('{botUsername}', client.user.username);
|
Act as a unique user in a Discord server. Engage in conversations naturally, considering the context and your surroundings.
|
||||||
|
Share your thoughts, opinions, and experiences as if you were a real person. Use emojis, slang, and a casual tone to blend in with other users.
|
||||||
|
Avoid sounding robotic or generic, and aim to create a distinct personality that others can relate to and enjoy interacting with.
|
||||||
|
`;
|
||||||
|
|
||||||
let messages = [{
|
let messages = [{
|
||||||
"role": "system",
|
"role": "system",
|
||||||
"content": prompt
|
"content": completionPrompt
|
||||||
}];
|
}];
|
||||||
|
|
||||||
let oldMessages;
|
let oldMessages;
|
||||||
if (conversations.has(message.author.id)) oldMessages = conversations.get(message.author.id);
|
if (conversations.has(message.author.id)) oldMessages = conversations.get(message.author.id);
|
||||||
if (oldMessages) {
|
if (oldMessages) {
|
||||||
// If there are old messages, check if they exceed token limit
|
|
||||||
while (func.tokenizer('gpt-3.5-turbo-0125', oldMessages).tokens >= 512) {
|
while (func.tokenizer('gpt-3.5-turbo-0125', oldMessages).tokens >= 512) {
|
||||||
let sliceLength = oldMessages.length * -0.5;
|
let sliceLength = oldMessages.length * -0.5;
|
||||||
if (sliceLength % 2 !== 0) sliceLength--;
|
if (sliceLength % 2 !== 0) sliceLength--;
|
||||||
oldMessages = oldMessages.slice(sliceLength);
|
oldMessages = oldMessages.slice(sliceLength);
|
||||||
// Update the conversation history in the map
|
|
||||||
conversations.set(message.author.id, oldMessages);
|
conversations.set(message.author.id, oldMessages);
|
||||||
}
|
}
|
||||||
// Concatenate old messages with the current message
|
|
||||||
messages = messages.concat(oldMessages);
|
messages = messages.concat(oldMessages);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -289,61 +295,36 @@ if (chatbot.State && chatbot.AllowedChannels.includes(message.channel.name) && !
|
||||||
});
|
});
|
||||||
|
|
||||||
openai.chat.completions.create({
|
openai.chat.completions.create({
|
||||||
|
|
||||||
model: 'gpt-3.5-turbo-0125',
|
model: 'gpt-3.5-turbo-0125',
|
||||||
messages: messages,
|
messages: messages,
|
||||||
max_tokens: func.tokenizer('gpt-3.5-turbo-0125', messages).maxTokens,
|
max_tokens: func.tokenizer('gpt-3.5-turbo-0125', messages).maxTokens,
|
||||||
temperature: settings.completion.temprature,
|
temperature: 0.8,
|
||||||
top_p: settings.completion.top_p,
|
top_p: 1,
|
||||||
frequency_penalty: settings.completion.frequency_penalty,
|
frequency_penalty: 0.5,
|
||||||
presence_penalty: settings.completion.presence_penalty,
|
presence_penalty: 0.5,
|
||||||
stream: true
|
stream: true
|
||||||
|
|
||||||
}).then(async (response) => {
|
}).then(async (response) => {
|
||||||
|
let fullAnswer = '';
|
||||||
// Initialize an array to hold all response parts
|
|
||||||
let responseParts = [];
|
|
||||||
|
|
||||||
for await (const part of response) {
|
for await (const part of response) {
|
||||||
// Accumulate response parts
|
fullAnswer += part.choices[0]?.delta?.content || '';
|
||||||
responseParts.push(part.choices[0]?.delta?.content || '');
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Combine all response parts into a single string
|
await message.channel.send(fullAnswer);
|
||||||
let fullAnswer = responseParts.join('');
|
|
||||||
|
|
||||||
// Trim the response content to fit within the maximum embed description length
|
|
||||||
if (fullAnswer.length > 4096) {
|
|
||||||
fullAnswer = fullAnswer.slice(0, 4093) + '...';
|
|
||||||
}
|
|
||||||
|
|
||||||
// Send the combined response as an embed
|
|
||||||
const embed = {
|
|
||||||
color: 0x0099ff,
|
|
||||||
title: 'Assisto',
|
|
||||||
description: fullAnswer
|
|
||||||
};
|
|
||||||
|
|
||||||
// Send the embed
|
|
||||||
await message.channel.send({ embeds: [embed] });
|
|
||||||
|
|
||||||
// Update the conversation history in the map with the new message
|
|
||||||
conversations.set(message.author.id, messages.concat([{ "role": "assistant", "content": fullAnswer }]));
|
conversations.set(message.author.id, messages.concat([{ "role": "assistant", "content": fullAnswer }]));
|
||||||
|
|
||||||
}).catch(async (error) => {
|
}).catch(async (error) => {
|
||||||
|
|
||||||
console.error(chalk.bold.redBright(error));
|
console.error(chalk.bold.redBright(error));
|
||||||
|
|
||||||
if (error.response) await message.reply({ content: error.response.error.message.length > 4000 ? error.response.error.message.substring(0, 3097) + "..." : error.response.error.message });
|
if (error.response) await message.reply(error.response.error.message.substring(0, 2000));
|
||||||
else if (error.message) await message.reply({ content: error.message.length > 4000 ? error.message.substring(0, 3097) + "..." : error.message });
|
else if (error.message) await message.reply(error.message.substring(0, 2000));
|
||||||
|
|
||||||
});
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Command Handler
|
// Command Handler
|
||||||
if (message.content.toLowerCase().startsWith(config.Prefix)) {
|
if (message.content.toLowerCase().startsWith(config.Prefix)) {
|
||||||
|
|
||||||
const neededPermissions = [
|
const neededPermissions = [
|
||||||
"ViewChannel",
|
"ViewChannel",
|
||||||
"SendMessages",
|
"SendMessages",
|
||||||
|
@ -358,15 +339,64 @@ if (chatbot.State && chatbot.AllowedChannels.includes(message.channel.name) && !
|
||||||
const command = client.MessageCommands.get(cmd) || client.MessageCommands.find(c => c.aliases && c.aliases.map(a => a.toLowerCase()).includes(cmd));
|
const command = client.MessageCommands.get(cmd) || client.MessageCommands.find(c => c.aliases && c.aliases.map(a => a.toLowerCase()).includes(cmd));
|
||||||
|
|
||||||
if (command) {
|
if (command) {
|
||||||
|
|
||||||
try {
|
try {
|
||||||
command.execute(client, message, args, cmd);
|
command.execute(client, message, args, cmd);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
console.error(chalk.bold.redBright(error));
|
console.error(chalk.bold.redBright(error));
|
||||||
};
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
};
|
// Function to handle direct messages
|
||||||
|
async function handleDirectMessage(message) {
|
||||||
|
if (conversations.has(message.author.id)) {
|
||||||
|
const oldMessages = conversations.get(message.author.id);
|
||||||
|
|
||||||
};
|
let messages = [{
|
||||||
|
"role": "system",
|
||||||
|
"content": completionPrompt
|
||||||
|
}];
|
||||||
|
|
||||||
};
|
messages = messages.concat(oldMessages);
|
||||||
|
|
||||||
|
messages.push({
|
||||||
|
"role": "user",
|
||||||
|
"content": message.content
|
||||||
|
});
|
||||||
|
|
||||||
|
openai.chat.completions.create({
|
||||||
|
model: 'gpt-3.5-turbo-0125',
|
||||||
|
messages: messages,
|
||||||
|
max_tokens: func.tokenizer('gpt-3.5-turbo-0125', messages).maxTokens,
|
||||||
|
temperature: 0.8,
|
||||||
|
top_p: 1,
|
||||||
|
frequency_penalty: 0.5,
|
||||||
|
presence_penalty: 0.5,
|
||||||
|
stream: true
|
||||||
|
}).then(async (response) => {
|
||||||
|
let fullAnswer = '';
|
||||||
|
|
||||||
|
for await (const part of response) {
|
||||||
|
fullAnswer += part.choices[0]?.delta?.content || '';
|
||||||
|
}
|
||||||
|
|
||||||
|
await message.author.send(fullAnswer);
|
||||||
|
|
||||||
|
conversations.set(message.author.id, messages.concat([{ "role": "assistant", "content": fullAnswer }]));
|
||||||
|
}).catch(async (error) => {
|
||||||
|
console.error(chalk.bold.redBright(error));
|
||||||
|
|
||||||
|
if (error.response) await message.author.send(error.response.error.message.substring(0, 2000));
|
||||||
|
else if (error.message) await message.author.send(error.message.substring(0, 2000));
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
await message.author.send("Hey there! What's up? Feel free to chat with me about anything!");
|
||||||
|
|
||||||
|
conversations.set(message.author.id, []);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Function to determine if the bot should respond based on its personality and surroundings
|
||||||
|
function shouldRespond(message) {
|
||||||
|
return true;
|
||||||
|
}}
|
Loading…
Reference in a new issue