diff --git a/js/llm.ts b/js/llm.ts index c8169b2..3e93508 100644 --- a/js/llm.ts +++ b/js/llm.ts @@ -22,7 +22,7 @@ export type LLMArgs = { temperature?: number; } & OpenAIAuth; -export const DEFAULT_MODEL = "gpt-4o"; +export const DEFAULT_MODEL = "gpt-4o-mini"; const PLAIN_RESPONSE_SCHEMA = { properties: { @@ -217,7 +217,7 @@ export function LLMClassifierFromTemplate({ const prompt = promptTemplate + "\n" + (useCoT ? COT_SUFFIX : NO_COT_SUFFIX); - let maxTokens = 512; + const maxTokens = 512; const messages: ChatCompletionMessageParam[] = [ { role: "user",