From f91f4dff7b2933a0213c439cfffaa2f4222d5948 Mon Sep 17 00:00:00 2001 From: bignutty <3515180-bignutty@users.noreply.gitlab.com> Date: Mon, 13 May 2024 15:30:29 +0200 Subject: [PATCH] reflect mono2 changes (c4e83892) --- commands/message/genai/imagen.js | 4 ++-- labscore/api/obelisk/endpoints.js | 2 ++ labscore/api/obelisk/index.js | 8 ++++++++ 3 files changed, 12 insertions(+), 2 deletions(-) diff --git a/commands/message/genai/imagen.js b/commands/message/genai/imagen.js index 68ae190..89f2688 100644 --- a/commands/message/genai/imagen.js +++ b/commands/message/genai/imagen.js @@ -1,4 +1,4 @@ -const { imagen } = require("../../../labscore/api/obelisk"); +const { GenerativeImagesModelsImagen } = require("../../../labscore/api/obelisk"); const { createEmbed } = require("../../../labscore/utils/embed"); const { editOrReply } = require("../../../labscore/utils/message"); @@ -41,7 +41,7 @@ module.exports = { await editOrReply(context, {embeds: loadingEmbeds}); - let res = await imagen(context, args.text); + let res = await GenerativeImagesModelsImagen(context, args.text); // Construct Embeds let files = []; diff --git a/labscore/api/obelisk/endpoints.js b/labscore/api/obelisk/endpoints.js index 63270b7..e5d2338 100644 --- a/labscore/api/obelisk/endpoints.js +++ b/labscore/api/obelisk/endpoints.js @@ -10,6 +10,8 @@ const ObeliskApi = Object.freeze({ LLM_PRIVATE_BARD: "/llm/v1/_private:bard", LLM_MODELS_GENERATE: "/llm/v1/generate", + GENERATIVEIMAGES_MODELS_IMAGEN: "/generativeimages/v1/models/imagen", + GOOGLE_BARD: "/parrot/v1/google:bard", GOOGLE_GEMINI_PRO: "/parrot/v1/google:gemini", GOOGLE_GEMINI_PRO_VISION: "/parrot/v1/google:geminiVision", diff --git a/labscore/api/obelisk/index.js b/labscore/api/obelisk/index.js index 3d0b8ae..30fbe02 100644 --- a/labscore/api/obelisk/index.js +++ b/labscore/api/obelisk/index.js @@ -56,6 +56,14 @@ module.exports.LlmModelsGenerate = async function(context, model, prompt, harmLe }) } +module.exports.GenerativeImagesModelsImagen = async function(context, prompt){ + return await request(ObeliskApi.GENERATIVEIMAGES_MODELS_IMAGEN, "POST", {}, { + image_prompt: prompt + }) +} + + + // GENERATIVEAI module.exports.bard = async function(context, input){ return await request(ObeliskApi.GOOGLE_BARD, "POST", {}, {