From ed955a74bed6991e1d775226855586e771d593b2 Mon Sep 17 00:00:00 2001 From: github-actions <41898282+github-actions[bot]@users.noreply.github.com> Date: Fri, 26 Jan 2024 07:08:34 +0000 Subject: [PATCH] =?UTF-8?q?=F0=9F=9A=A8=20Commit=20Build=20Artifact=20from?= =?UTF-8?q?=20GitHub=20Actions?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- dist/main_bun.mjs | 35 ++++++++++++++++++++++++-------- dist/main_cloudflare-workers.mjs | 35 ++++++++++++++++++++++++-------- dist/main_deno.mjs | 35 ++++++++++++++++++++++++-------- dist/main_node.mjs | 35 ++++++++++++++++++++++++-------- 4 files changed, 108 insertions(+), 32 deletions(-) diff --git a/dist/main_bun.mjs b/dist/main_bun.mjs index 7420558..02f41d8 100644 --- a/dist/main_bun.mjs +++ b/dist/main_bun.mjs @@ -1331,14 +1331,19 @@ var LogLevel = /* @__PURE__ */ ((LogLevel2) => { LogLevel2[LogLevel2["debug"] = 7] = "debug"; return LogLevel2; })(LogLevel || {}); +var currentlevel = 3 /* error */; function gen_logger(id) { return mapValues(LogLevel, (value, name) => { return (msg) => { - outFunc(name, value, `${id} ${msg}`); + outFunc(name, value, `${id} ${JSON.stringify(msg)}`); }; }); } -function outFunc(_levelName, levelValue, _msg) { +function outFunc(levelName, levelValue, msg) { + if (levelValue > currentlevel) { + return; + } + console.log(`${Date.now().toLocaleString()} ${levelName} ${msg}`); } function mapValues(obj, fn) { return Object.fromEntries( @@ -1976,7 +1981,16 @@ function genModel(genAi, req) { maxOutputTokens: req.max_tokens ?? void 0, temperature: req.temperature ?? void 0, topP: req.top_p ?? void 0 - } + }, + safetySettings: [ + HarmCategory.HARM_CATEGORY_HATE_SPEECH, + HarmCategory.HARM_CATEGORY_SEXUALLY_EXPLICIT, + HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, + HarmCategory.HARM_CATEGORY_HARASSMENT + ].map((category) => ({ + category, + threshold: HarmBlockThreshold.BLOCK_NONE + })) }); return model; } @@ -1987,8 +2001,12 @@ var nonStreamingChatProxyHandler = async (c, req, genAi) => { const model = genModel(genAi, req); const geminiResp = await model.generateContent({ contents: openAiMessageToGeminiMessage(req.messages) - }).then((it) => it.response.text()).catch((err) => err?.message ?? err.toString()); - log2.debug(JSON.stringify(geminiResp)); + }).then((it) => it.response.text()).catch((err) => { + log2.error(req); + log2.error(err); + return err?.message ?? err.toString(); + }); + log2.debug(geminiResp); const resp = { id: "chatcmpl-abc123", object: "chat.completion", @@ -2063,12 +2081,13 @@ var streamingChatProxyHandler = async (c, req, genAi) => { data: JSON.stringify(genOpenAiResp("", true)) }); const geminiResult = (await response).text(); - log2.info(JSON.stringify(geminiResult)); + log2.info(geminiResult); }).catch(async (e) => { await sseStream.writeSSE({ data: JSON.stringify(genOpenAiResp(e.toString(), true)) }); - log2.info(e); + log2.error(req); + log2.error(e); }); await sseStream.writeSSE({ data: "[DONE]" }); await sseStream.close(); @@ -2079,7 +2098,7 @@ var streamingChatProxyHandler = async (c, req, genAi) => { var chatProxyHandler = async (c) => { const log2 = c.var.log; const req = await c.req.json(); - log2.debug(JSON.stringify(req)); + log2.debug(req); const headers = c.req.header(); const apiKey = getToken(headers); if (apiKey == null) { diff --git a/dist/main_cloudflare-workers.mjs b/dist/main_cloudflare-workers.mjs index 7a42554..f5dc424 100644 --- a/dist/main_cloudflare-workers.mjs +++ b/dist/main_cloudflare-workers.mjs @@ -1331,14 +1331,19 @@ var LogLevel = /* @__PURE__ */ ((LogLevel2) => { LogLevel2[LogLevel2["debug"] = 7] = "debug"; return LogLevel2; })(LogLevel || {}); +var currentlevel = 3 /* error */; function gen_logger(id) { return mapValues(LogLevel, (value, name) => { return (msg) => { - outFunc(name, value, `${id} ${msg}`); + outFunc(name, value, `${id} ${JSON.stringify(msg)}`); }; }); } -function outFunc(_levelName, levelValue, _msg) { +function outFunc(levelName, levelValue, msg) { + if (levelValue > currentlevel) { + return; + } + console.log(`${Date.now().toLocaleString()} ${levelName} ${msg}`); } function mapValues(obj, fn) { return Object.fromEntries( @@ -1976,7 +1981,16 @@ function genModel(genAi, req) { maxOutputTokens: req.max_tokens ?? void 0, temperature: req.temperature ?? void 0, topP: req.top_p ?? void 0 - } + }, + safetySettings: [ + HarmCategory.HARM_CATEGORY_HATE_SPEECH, + HarmCategory.HARM_CATEGORY_SEXUALLY_EXPLICIT, + HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, + HarmCategory.HARM_CATEGORY_HARASSMENT + ].map((category) => ({ + category, + threshold: HarmBlockThreshold.BLOCK_NONE + })) }); return model; } @@ -1987,8 +2001,12 @@ var nonStreamingChatProxyHandler = async (c, req, genAi) => { const model = genModel(genAi, req); const geminiResp = await model.generateContent({ contents: openAiMessageToGeminiMessage(req.messages) - }).then((it) => it.response.text()).catch((err) => err?.message ?? err.toString()); - log2.debug(JSON.stringify(geminiResp)); + }).then((it) => it.response.text()).catch((err) => { + log2.error(req); + log2.error(err); + return err?.message ?? err.toString(); + }); + log2.debug(geminiResp); const resp = { id: "chatcmpl-abc123", object: "chat.completion", @@ -2063,12 +2081,13 @@ var streamingChatProxyHandler = async (c, req, genAi) => { data: JSON.stringify(genOpenAiResp("", true)) }); const geminiResult = (await response).text(); - log2.info(JSON.stringify(geminiResult)); + log2.info(geminiResult); }).catch(async (e) => { await sseStream.writeSSE({ data: JSON.stringify(genOpenAiResp(e.toString(), true)) }); - log2.info(e); + log2.error(req); + log2.error(e); }); await sseStream.writeSSE({ data: "[DONE]" }); await sseStream.close(); @@ -2079,7 +2098,7 @@ var streamingChatProxyHandler = async (c, req, genAi) => { var chatProxyHandler = async (c) => { const log2 = c.var.log; const req = await c.req.json(); - log2.debug(JSON.stringify(req)); + log2.debug(req); const headers = c.req.header(); const apiKey = getToken(headers); if (apiKey == null) { diff --git a/dist/main_deno.mjs b/dist/main_deno.mjs index 824ddb7..59d09a9 100644 --- a/dist/main_deno.mjs +++ b/dist/main_deno.mjs @@ -1331,14 +1331,19 @@ var LogLevel = /* @__PURE__ */ ((LogLevel2) => { LogLevel2[LogLevel2["debug"] = 7] = "debug"; return LogLevel2; })(LogLevel || {}); +var currentlevel = 3 /* error */; function gen_logger(id) { return mapValues(LogLevel, (value, name) => { return (msg) => { - outFunc(name, value, `${id} ${msg}`); + outFunc(name, value, `${id} ${JSON.stringify(msg)}`); }; }); } -function outFunc(_levelName, levelValue, _msg) { +function outFunc(levelName, levelValue, msg) { + if (levelValue > currentlevel) { + return; + } + console.log(`${Date.now().toLocaleString()} ${levelName} ${msg}`); } function mapValues(obj, fn) { return Object.fromEntries( @@ -1976,7 +1981,16 @@ function genModel(genAi, req) { maxOutputTokens: req.max_tokens ?? void 0, temperature: req.temperature ?? void 0, topP: req.top_p ?? void 0 - } + }, + safetySettings: [ + HarmCategory.HARM_CATEGORY_HATE_SPEECH, + HarmCategory.HARM_CATEGORY_SEXUALLY_EXPLICIT, + HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, + HarmCategory.HARM_CATEGORY_HARASSMENT + ].map((category) => ({ + category, + threshold: HarmBlockThreshold.BLOCK_NONE + })) }); return model; } @@ -1987,8 +2001,12 @@ var nonStreamingChatProxyHandler = async (c, req, genAi) => { const model = genModel(genAi, req); const geminiResp = await model.generateContent({ contents: openAiMessageToGeminiMessage(req.messages) - }).then((it) => it.response.text()).catch((err) => err?.message ?? err.toString()); - log2.debug(JSON.stringify(geminiResp)); + }).then((it) => it.response.text()).catch((err) => { + log2.error(req); + log2.error(err); + return err?.message ?? err.toString(); + }); + log2.debug(geminiResp); const resp = { id: "chatcmpl-abc123", object: "chat.completion", @@ -2063,12 +2081,13 @@ var streamingChatProxyHandler = async (c, req, genAi) => { data: JSON.stringify(genOpenAiResp("", true)) }); const geminiResult = (await response).text(); - log2.info(JSON.stringify(geminiResult)); + log2.info(geminiResult); }).catch(async (e) => { await sseStream.writeSSE({ data: JSON.stringify(genOpenAiResp(e.toString(), true)) }); - log2.info(e); + log2.error(req); + log2.error(e); }); await sseStream.writeSSE({ data: "[DONE]" }); await sseStream.close(); @@ -2079,7 +2098,7 @@ var streamingChatProxyHandler = async (c, req, genAi) => { var chatProxyHandler = async (c) => { const log2 = c.var.log; const req = await c.req.json(); - log2.debug(JSON.stringify(req)); + log2.debug(req); const headers = c.req.header(); const apiKey = getToken(headers); if (apiKey == null) { diff --git a/dist/main_node.mjs b/dist/main_node.mjs index 1044f0d..ccb6c28 100644 --- a/dist/main_node.mjs +++ b/dist/main_node.mjs @@ -1642,14 +1642,19 @@ var LogLevel = /* @__PURE__ */ ((LogLevel2) => { LogLevel2[LogLevel2["debug"] = 7] = "debug"; return LogLevel2; })(LogLevel || {}); +var currentlevel = 3 /* error */; function gen_logger(id) { return mapValues(LogLevel, (value, name) => { return (msg) => { - outFunc(name, value, `${id} ${msg}`); + outFunc(name, value, `${id} ${JSON.stringify(msg)}`); }; }); } -function outFunc(_levelName, levelValue, _msg) { +function outFunc(levelName, levelValue, msg) { + if (levelValue > currentlevel) { + return; + } + console.log(`${Date.now().toLocaleString()} ${levelName} ${msg}`); } function mapValues(obj, fn) { return Object.fromEntries( @@ -2287,7 +2292,16 @@ function genModel(genAi, req) { maxOutputTokens: req.max_tokens ?? void 0, temperature: req.temperature ?? void 0, topP: req.top_p ?? void 0 - } + }, + safetySettings: [ + HarmCategory.HARM_CATEGORY_HATE_SPEECH, + HarmCategory.HARM_CATEGORY_SEXUALLY_EXPLICIT, + HarmCategory.HARM_CATEGORY_DANGEROUS_CONTENT, + HarmCategory.HARM_CATEGORY_HARASSMENT + ].map((category) => ({ + category, + threshold: HarmBlockThreshold.BLOCK_NONE + })) }); return model; } @@ -2298,8 +2312,12 @@ var nonStreamingChatProxyHandler = async (c, req, genAi) => { const model = genModel(genAi, req); const geminiResp = await model.generateContent({ contents: openAiMessageToGeminiMessage(req.messages) - }).then((it) => it.response.text()).catch((err) => err?.message ?? err.toString()); - log2.debug(JSON.stringify(geminiResp)); + }).then((it) => it.response.text()).catch((err) => { + log2.error(req); + log2.error(err); + return err?.message ?? err.toString(); + }); + log2.debug(geminiResp); const resp = { id: "chatcmpl-abc123", object: "chat.completion", @@ -2374,12 +2392,13 @@ var streamingChatProxyHandler = async (c, req, genAi) => { data: JSON.stringify(genOpenAiResp("", true)) }); const geminiResult = (await response).text(); - log2.info(JSON.stringify(geminiResult)); + log2.info(geminiResult); }).catch(async (e) => { await sseStream.writeSSE({ data: JSON.stringify(genOpenAiResp(e.toString(), true)) }); - log2.info(e); + log2.error(req); + log2.error(e); }); await sseStream.writeSSE({ data: "[DONE]" }); await sseStream.close(); @@ -2390,7 +2409,7 @@ var streamingChatProxyHandler = async (c, req, genAi) => { var chatProxyHandler = async (c) => { const log2 = c.var.log; const req = await c.req.json(); - log2.debug(JSON.stringify(req)); + log2.debug(req); const headers = c.req.header(); const apiKey = getToken(headers); if (apiKey == null) {