Perplexity AI (pplx-api)
API Key
# env variable
os.environ['PERPLEXITYAI_API_KEY']
Sample Usage
from litellm import completion
import os
os.environ['PERPLEXITYAI_API_KEY'] = ""
response = completion(
model="perplexity/mistral-7b-instruct",
messages=messages
)
print(response)
Sample Usage - Streaming
from litellm import completion
import os
os.environ['PERPLEXITYAI_API_KEY'] = ""
response = completion(
model="perplexity/mistral-7b-instruct",
messages=messages,
stream=True
)
for chunk in response:
print(chunk)
Supported Models
All models listed here https://docs.perplexity.ai/docs/model-cards are supported
Model Name | Function Call |
---|---|
pplx-7b-chat | completion(model="perplexity/pplx-7b-chat", messages) |
pplx-70b-chat | completion(model="perplexity/pplx-70b-chat", messages) |
pplx-7b-online | completion(model="perplexity/pplx-7b-online", messages) |
pplx-70b-online | completion(model="perplexity/pplx-70b-online", messages) |
codellama-34b-instruct | completion(model="perplexity/codellama-34b-instruct", messages) |
llama-2-13b-chat | completion(model="perplexity/llama-2-13b-chat", messages) |
llama-2-70b-chat | completion(model="perplexity/llama-2-70b-chat", messages) |
mistral-7b-instruct | completion(model="perplexity/mistral-7b-instruct", messages) |
openhermes-2-mistral-7b | completion(model="perplexity/openhermes-2-mistral-7b", messages) |
openhermes-2.5-mistral-7b | completion(model="perplexity/openhermes-2.5-mistral-7b", messages) |
pplx-7b-chat-alpha | completion(model="perplexity/pplx-7b-chat-alpha", messages) |
pplx-70b-chat-alpha | completion(model="perplexity/pplx-70b-chat-alpha", messages) |