Token.js
GitHubDiscordFeedback
  • Token.js
  • Providers
    • AI21
    • Anthropic
    • Bedrock
    • Cohere
    • Gemini
    • Groq
    • Mistral
    • OpenAI
    • OpenRouter
    • Perplexity
    • OpenAI Compatible
  • Contact Us
  • Contributing
Powered by GitBook
On this page
  • Usage
  • Supported Models
  • Legend
  • Additional Resources

Was this helpful?

Edit on GitHub
  1. Providers

Perplexity

PreviousOpenRouterNextOpenAI Compatible

Last updated 3 months ago

Was this helpful?

Usage

.env
PERPLEXITY_API_KEY=
import { TokenJS } from 'token.js'

// Create the Token.js client
const tokenjs = new TokenJS()

async function main() {
  // Create a model response
  const completion = await tokenjs.chat.completions.create({
    // Specify the provider and model
    provider: 'perplexity',
    model: 'llama-3-70b-instruct',
    // Define your message
    messages: [
      {
        role: 'user',
        content: 'Hello!',
      },
    ],
  })
  console.log(completion.choices[0])
}
main()

Supported Models

Model
Chat Completion
Streaming
JSON Output
Image Input
Function Calling
N > 1

llama-3.1-sonar-small-128k-online

✅

✅

➖

➖

➖

➖

llama-3.1-sonar-large-128k-online

✅

✅

➖

➖

➖

➖

llama-3.1-sonar-huge-128k-online

✅

✅

➖

➖

➖

➖

Legend

Symbol
Description

Supported by Token.js

Not supported by the LLM provider, so Token.js cannot support it

Additional Resources

Get a Perplexity API key
Perplexity Documentation
✅
➖