Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
80 changes: 16 additions & 64 deletions lib/chat-setting-limits.ts
Original file line number Diff line number Diff line change
Expand Up @@ -158,83 +158,35 @@ export const CHAT_SETTING_LIMITS: Record<LLMID, ChatSettingLimits> = {
MAX_CONTEXT_LENGTH: 128000
},

// PERPLEXITY MODELS
"pplx-7b-online": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.99,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 4096
},
"pplx-70b-online": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.99,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 4096
},
"pplx-7b-chat": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 8192
},
"pplx-70b-chat": {
// PERPLEXITY MODELS (UPDATED 3/16/26)
"sonar-reasoning-pro": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 4096
},
"mixtral-8x7b-instruct": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 16384,
MAX_CONTEXT_LENGTH: 16384
},
"mistral-7b-instruct": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 16384,
MAX_CONTEXT_LENGTH: 16384
},
"llama-2-70b-chat": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 2.0,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 4096
},
"codellama-34b-instruct": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 4096,
MAX_CONTEXT_LENGTH: 16384
},
"codellama-70b-instruct": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 16384,
MAX_CONTEXT_LENGTH: 16384
MAX_TOKEN_OUTPUT_LENGTH: 8192,
MAX_CONTEXT_LENGTH: 128000
},
"sonar-small-chat": {
"sonar-reasoning": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 16384,
MAX_CONTEXT_LENGTH: 16384
MAX_TOKEN_OUTPUT_LENGTH: 8192,
MAX_CONTEXT_LENGTH: 128000
},
"sonar-small-online": {
"sonar-pro": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 12000,
MAX_CONTEXT_LENGTH: 12000
MAX_TOKEN_OUTPUT_LENGTH: 8192,
MAX_CONTEXT_LENGTH: 200000
},
"sonar-medium-chat": {
"sonar": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 16384,
MAX_CONTEXT_LENGTH: 16384
MAX_TOKEN_OUTPUT_LENGTH: 8192,
MAX_CONTEXT_LENGTH: 128000
},
"sonar-medium-online": {
"sonar-deep-research": {
MIN_TEMPERATURE: 0.0,
MAX_TEMPERATURE: 1.0,
MAX_TOKEN_OUTPUT_LENGTH: 12000,
MAX_CONTEXT_LENGTH: 12000
MAX_TOKEN_OUTPUT_LENGTH: 8192,
MAX_CONTEXT_LENGTH: 128000
}
}
127 changes: 67 additions & 60 deletions lib/models/llm/perplexity-llm-list.ts
Original file line number Diff line number Diff line change
Expand Up @@ -3,86 +3,93 @@ import { LLM } from "@/types"
const PERPLEXITY_PLATORM_LINK =
"https://docs.perplexity.ai/docs/getting-started"

// Perplexity Models (UPDATED 2/25/24) -----------------------------
// Model Deprecation Notice
// Please note that on March 15, the pplx-70b-chat, pplx-70b-online, llama-2-70b-chat, and codellama-34b-instruct models will no longer be available through the Perplexity API.
// Perplexity Models (UPDATED 3/16/26) -----------------------------
// Documentation: https://docs.perplexity.ai/guides/model-cards

// Mixtral 8x7B Instruct (UPDATED 1/31/24)
const MIXTRAL_8X7B_INSTRUCT: LLM = {
modelId: "mixtral-8x7b-instruct",
modelName: "Mixtral 8x7B Instruct",
// Sonar Reasoning Pro - Advanced search with Chain of Thought reasoning
const SONAR_REASONING_PRO: LLM = {
modelId: "sonar-reasoning-pro",
modelName: "Sonar Reasoning Pro",
provider: "perplexity",
hostedId: "mixtral-8x7b-instruct",
hostedId: "sonar-reasoning-pro",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
imageInput: false,
pricing: {
currency: "USD",
unit: "1M tokens",
inputCost: 2,
outputCost: 8
}
}

// Mistral 7B Instruct (UPDATED 1/31/24)
const MISTRAL_7B_INSTRUCT: LLM = {
modelId: "mistral-7b-instruct",
modelName: "Mistral 7B Instruct",
// Sonar Reasoning - Cost-effective reasoning model
const SONAR_REASONING: LLM = {
modelId: "sonar-reasoning",
modelName: "Sonar Reasoning",
provider: "perplexity",
hostedId: "mistral-7b-instruct",
hostedId: "sonar-reasoning",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
imageInput: false,
pricing: {
currency: "USD",
unit: "1M tokens",
inputCost: 1,
outputCost: 5
}
}

// CodeLlama 70B Instruct (UPDATED 1/31/24)
const CODELLAMA_70B_INSTRUCT: LLM = {
modelId: "codellama-70b-instruct",
modelName: "CodeLlama 70B Instruct",
// Sonar Pro - Advanced search for complex queries
const SONAR_PRO: LLM = {
modelId: "sonar-pro",
modelName: "Sonar Pro",
provider: "perplexity",
hostedId: "codellama-70b-instruct",
hostedId: "sonar-pro",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
imageInput: false,
pricing: {
currency: "USD",
unit: "1M tokens",
inputCost: 3,
outputCost: 15
}
}

// Sonar Small Chat (UPDATED 2/25/24)
const PERPLEXITY_SONAR_SMALL_CHAT_7B: LLM = {
modelId: "sonar-small-chat",
modelName: "Sonar Small Chat",
// Sonar - Lightweight model for simple tasks
const SONAR: LLM = {
modelId: "sonar",
modelName: "Sonar",
provider: "perplexity",
hostedId: "sonar-small-chat",
hostedId: "sonar",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
imageInput: false,
pricing: {
currency: "USD",
unit: "1M tokens",
inputCost: 1,
outputCost: 1
}
}

// Sonar Small Online (UPDATED 2/25/24)
const PERPLEXITY_SONAR_SMALL_ONLINE_7B: LLM = {
modelId: "sonar-small-online",
modelName: "Sonar Small Online",
// Sonar Deep Research - Expert-level research model
const SONAR_DEEP_RESEARCH: LLM = {
modelId: "sonar-deep-research",
modelName: "Sonar Deep Research",
provider: "perplexity",
hostedId: "sonar-small-online",
hostedId: "sonar-deep-research",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
}

// Sonar Medium Chat (UPDATED 2/25/24)
const PERPLEXITY_SONAR_MEDIUM_CHAT_8x7B: LLM = {
modelId: "sonar-medium-chat",
modelName: "Sonar Medium Chat",
provider: "perplexity",
hostedId: "sonar-medium-chat",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
}

// Sonar Medium Online (UPDATED 2/25/24)
const PERPLEXITY_SONAR_MEDIUM_ONLINE_8x7B: LLM = {
modelId: "sonar-medium-online",
modelName: "Sonar Medium Online",
provider: "perplexity",
hostedId: "sonar-medium-online",
platformLink: PERPLEXITY_PLATORM_LINK,
imageInput: false
imageInput: false,
pricing: {
currency: "USD",
unit: "1M tokens",
inputCost: 2,
outputCost: 8
}
}

export const PERPLEXITY_LLM_LIST: LLM[] = [
MIXTRAL_8X7B_INSTRUCT,
MISTRAL_7B_INSTRUCT,
CODELLAMA_70B_INSTRUCT,
PERPLEXITY_SONAR_SMALL_CHAT_7B,
PERPLEXITY_SONAR_SMALL_ONLINE_7B,
PERPLEXITY_SONAR_MEDIUM_CHAT_8x7B,
PERPLEXITY_SONAR_MEDIUM_ONLINE_8x7B
SONAR_REASONING_PRO,
SONAR_REASONING,
SONAR_PRO,
SONAR,
SONAR_DEEP_RESEARCH
]
20 changes: 6 additions & 14 deletions types/llms.ts
Original file line number Diff line number Diff line change
Expand Up @@ -45,21 +45,13 @@ export type GroqLLMID =
| "mixtral-8x7b-32768" // Mixtral-8x7b
| "gemma-7b-it" // Gemma-7b IT

// Perplexity Models (UPDATED 1/31/24)
// Perplexity Models (UPDATED 3/16/26)
export type PerplexityLLMID =
| "pplx-7b-online" // Perplexity Online 7B
| "pplx-70b-online" // Perplexity Online 70B
| "pplx-7b-chat" // Perplexity Chat 7B
| "pplx-70b-chat" // Perplexity Chat 70B
| "mixtral-8x7b-instruct" // Mixtral 8x7B Instruct
| "mistral-7b-instruct" // Mistral 7B Instruct
| "llama-2-70b-chat" // Llama2 70B Chat
| "codellama-34b-instruct" // CodeLlama 34B Instruct
| "codellama-70b-instruct" // CodeLlama 70B Instruct
| "sonar-small-chat" // Sonar Small Chat
| "sonar-small-online" // Sonar Small Online
| "sonar-medium-chat" // Sonar Medium Chat
| "sonar-medium-online" // Sonar Medium Online
| "sonar-reasoning-pro" // Sonar Reasoning Pro
| "sonar-reasoning" // Sonar Reasoning
| "sonar-pro" // Sonar Pro
| "sonar" // Sonar
| "sonar-deep-research" // Sonar Deep Research

export interface LLM {
modelId: LLMID
Expand Down