diff --git a/commands/interaction/subcommands/gpt/alpaca.js b/commands/interaction/subcommands/gpt/alpaca.js index db3c27a..3abd7d4 100644 --- a/commands/interaction/subcommands/gpt/alpaca.js +++ b/commands/interaction/subcommands/gpt/alpaca.js @@ -26,9 +26,7 @@ module.exports = { try{ let s = Date.now() await context.respond({data: {}, type: InteractionCallbackTypes.DEFERRED_CHANNEL_MESSAGE_WITH_SOURCE}) - await context.editOrRespond({ - embeds: [createEmbed("loading_ai", context)] - }) + let res = await superagent.get(`${process.env.AI_SERVER}/gpt`) .query({ model: MODEL, diff --git a/commands/interaction/subcommands/gpt/chatgpt.js b/commands/interaction/subcommands/gpt/chatgpt.js index a4062d1..3ee1ea2 100644 --- a/commands/interaction/subcommands/gpt/chatgpt.js +++ b/commands/interaction/subcommands/gpt/chatgpt.js @@ -25,9 +25,7 @@ module.exports = { try{ let s = Date.now() await context.respond({data: {}, type: InteractionCallbackTypes.DEFERRED_CHANNEL_MESSAGE_WITH_SOURCE}) - await context.editOrRespond({ - embeds: [createEmbed("loading_ai", context)] - }) + let res = await superagent.get(`${process.env.AI_SERVER}/gpt`) .query({ model: MODEL, diff --git a/commands/interaction/subcommands/gpt/claude-instant.js b/commands/interaction/subcommands/gpt/claude-instant.js index 9e6d73b..7b6c0fc 100644 --- a/commands/interaction/subcommands/gpt/claude-instant.js +++ b/commands/interaction/subcommands/gpt/claude-instant.js @@ -25,9 +25,7 @@ module.exports = { try{ let s = Date.now() await context.respond({data: {}, type: InteractionCallbackTypes.DEFERRED_CHANNEL_MESSAGE_WITH_SOURCE}) - await context.editOrRespond({ - embeds: [createEmbed("loading_ai", context)] - }) + let res = await superagent.get(`${process.env.AI_SERVER}/gpt`) .query({ model: MODEL, diff --git a/commands/interaction/subcommands/gpt/claude.js b/commands/interaction/subcommands/gpt/claude.js index f4b71f6..eea4efc 100644 --- a/commands/interaction/subcommands/gpt/claude.js +++ b/commands/interaction/subcommands/gpt/claude.js @@ -25,9 +25,7 @@ module.exports = { try{ let s = Date.now() await context.respond({data: {}, type: InteractionCallbackTypes.DEFERRED_CHANNEL_MESSAGE_WITH_SOURCE}) - await context.editOrRespond({ - embeds: [createEmbed("loading_ai", context)] - }) + let res = await superagent.get(`${process.env.AI_SERVER}/gpt`) .query({ model: MODEL, diff --git a/commands/interaction/subcommands/gpt/davinci3.js b/commands/interaction/subcommands/gpt/davinci3.js index 80d6630..203f38c 100644 --- a/commands/interaction/subcommands/gpt/davinci3.js +++ b/commands/interaction/subcommands/gpt/davinci3.js @@ -25,9 +25,7 @@ module.exports = { try{ let s = Date.now() await context.respond({data: {}, type: InteractionCallbackTypes.DEFERRED_CHANNEL_MESSAGE_WITH_SOURCE}) - await context.editOrRespond({ - embeds: [createEmbed("loading_ai", context)] - }) + let res = await superagent.get(`${process.env.AI_SERVER}/gpt`) .query({ model: MODEL, diff --git a/labscore/utils/embed.js b/labscore/utils/embed.js index 40a5867..235902c 100644 --- a/labscore/utils/embed.js +++ b/labscore/utils/embed.js @@ -69,15 +69,6 @@ const embedTypes = Object.freeze({ }, color: COLORS.embed } - }, - "loading_ai": (context) => { - return { - author: { - iconUrl: STATIC_ICONS.loading_ai, - name: `Generating...` - }, - color: COLORS.brand - } } }) diff --git a/labscore/utils/statics.js b/labscore/utils/statics.js index a9100d5..1f534be 100644 --- a/labscore/utils/statics.js +++ b/labscore/utils/statics.js @@ -82,10 +82,6 @@ const Statics = Object.freeze({ file: "icons/core/ico_notice_loading.gif", revision: 0 }, - loading_ai: { - file: "icons/ai/ico_ai_generating.gif", - revision: 0 - }, warning: { file: "icons/core/ico_notice_warning.png", revision: 0 @@ -124,6 +120,5 @@ module.exports.STATIC_ICONS = Object.freeze({ adult: staticAsset(Statics.icons.adult), error: staticAsset(Statics.icons.error), loading: staticAsset(Statics.icons.loading), - loading_ai: staticAsset(Statics.icons.loading_ai), warning: staticAsset(Statics.icons.warning) }) \ No newline at end of file