include prompt in token count (#104)
This commit is contained in:
parent
7ce2d5ec2c
commit
1a4b4401ee
|
@ -7,7 +7,7 @@ import { init, Tiktoken } from "@dqbd/tiktoken/lite/init";
|
||||||
import wasm from "../../node_modules/@dqbd/tiktoken/lite/tiktoken_bg.wasm?module";
|
import wasm from "../../node_modules/@dqbd/tiktoken/lite/tiktoken_bg.wasm?module";
|
||||||
|
|
||||||
export const config = {
|
export const config = {
|
||||||
runtime: "edge"
|
runtime: "edge",
|
||||||
};
|
};
|
||||||
|
|
||||||
const handler = async (req: Request): Promise<Response> => {
|
const handler = async (req: Request): Promise<Response> => {
|
||||||
|
@ -15,10 +15,22 @@ const handler = async (req: Request): Promise<Response> => {
|
||||||
const { model, messages, key, prompt } = (await req.json()) as ChatBody;
|
const { model, messages, key, prompt } = (await req.json()) as ChatBody;
|
||||||
|
|
||||||
await init((imports) => WebAssembly.instantiate(wasm, imports));
|
await init((imports) => WebAssembly.instantiate(wasm, imports));
|
||||||
const encoding = new Tiktoken(tiktokenModel.bpe_ranks, tiktokenModel.special_tokens, tiktokenModel.pat_str);
|
const encoding = new Tiktoken(
|
||||||
|
tiktokenModel.bpe_ranks,
|
||||||
|
tiktokenModel.special_tokens,
|
||||||
|
tiktokenModel.pat_str
|
||||||
|
);
|
||||||
|
|
||||||
const tokenLimit = model.id === OpenAIModelID.GPT_4 ? 6000 : 3000;
|
const tokenLimit = model.id === OpenAIModelID.GPT_4 ? 6000 : 3000;
|
||||||
let tokenCount = 0;
|
|
||||||
|
let promptToSend = prompt;
|
||||||
|
if (!promptToSend) {
|
||||||
|
promptToSend = DEFAULT_SYSTEM_PROMPT;
|
||||||
|
}
|
||||||
|
|
||||||
|
const prompt_tokens = encoding.encode(promptToSend);
|
||||||
|
|
||||||
|
let tokenCount = prompt_tokens.length;
|
||||||
let messagesToSend: Message[] = [];
|
let messagesToSend: Message[] = [];
|
||||||
|
|
||||||
for (let i = messages.length - 1; i >= 0; i--) {
|
for (let i = messages.length - 1; i >= 0; i--) {
|
||||||
|
@ -34,11 +46,6 @@ const handler = async (req: Request): Promise<Response> => {
|
||||||
|
|
||||||
encoding.free();
|
encoding.free();
|
||||||
|
|
||||||
let promptToSend = prompt;
|
|
||||||
if (!promptToSend) {
|
|
||||||
promptToSend = DEFAULT_SYSTEM_PROMPT;
|
|
||||||
}
|
|
||||||
|
|
||||||
const stream = await OpenAIStream(model, promptToSend, key, messagesToSend);
|
const stream = await OpenAIStream(model, promptToSend, key, messagesToSend);
|
||||||
|
|
||||||
return new Response(stream);
|
return new Response(stream);
|
||||||
|
|
Loading…
Reference in New Issue