Perplexity
Usage
PERPLEXITY_API_KEY=
import { TokenJS } from 'token.js'
// Create the Token.js client
const tokenjs = new TokenJS()
async function main() {
// Create a model response
const completion = await tokenjs.chat.completions.create({
// Specify the provider and model
provider: 'perplexity',
model: 'llama-3-70b-instruct',
// Define your message
messages: [
{
role: 'user',
content: 'Hello!',
},
],
})
console.log(completion.choices[0])
}
main()
Supported Models
Model
Chat Completion
Streaming
JSON Output
Image Input
Function Calling
N > 1
llama-3.1-sonar-small-128k-online
✅
✅
➖
➖
➖
➖
llama-3.1-sonar-large-128k-online
✅
✅
➖
➖
➖
➖
llama-3.1-sonar-huge-128k-online
✅
✅
➖
➖
➖
➖
Legend
Symbol
Description
✅
Supported by Token.js
➖
Not supported by the LLM provider, so Token.js cannot support it
Additional Resources
Last updated
Was this helpful?