Overview

Julep leverages LiteLLM to seamlessly connect you to a wide array of Language Models (LLMs). This integration offers incredible flexibility, allowing you to tap into models from various providers with a straightforward, unified interface.

With our unified API, switching between different providers is a breeze, ensuring you maintain consistent functionality across the board.

Available Models

While we provide API keys for quick testing and development, you’ll need to use your own API keys when deploying to production. This ensures you have full control over your usage and billing.

Looking for top-notch quality? Our curated selection of models delivers excellent outputs for all your use cases.

Anthropic

Here are the Anthropic models supported by Julep:

Model NameContext WindowBest For
claude-3-opus200K tokensComplex reasoning, analysis
claude-3-sonnet200K tokensGeneral purpose tasks
claude-3-haiku200K tokensQuick responses
claude-3.5-sonnet200K tokensImproved reasoning
claude-3.5-sonnet-20240620200K tokensEnhanced reasoning capabilities
claude-3.5-sonnet-20241022200K tokensLatest improvements

Google

Here are the Google models supported by Julep:

Model NameContext WindowBest For
gemini-1.5-pro1M tokensComplex tasks
gemini-1.5-pro-latest1M tokensCutting-edge performance

OpenAI

Here are the OpenAI models supported by Julep:

Model NameContext WindowBest For
gpt-4-turbo128K tokensAdvanced reasoning
gpt-4o128K tokensBalanced performance
o1-mini¹128K tokensQuick tasks
o1-preview¹128K tokensTesting features
o1¹128K tokensGeneral tasks

¹ Heads up: The O1 series models are temporarily unavailable due to some OpenAI service issues. We’re working to get them back online soon!

Groq

Here are the Groq models supported by Julep:

Model NameContext WindowBest For
llama-3.1-70b8K tokensLong-form content
llama-3.1-8b8K tokensQuick processing

OpenRouter

Here are the OpenRouter models supported by Julep:

Model NameContext WindowBest For
mistral-large-2411128K tokensHigh performance
qwen-2.5-72b-instruct131K tokensComplex instructions
eva-llama-3.33-70b128K tokensCreative tasks
l3.1-euryale-70b128K tokensCreative tasks
l3.3-euryale-70b8K tokensCreative tasks
magnum-v4-72b8K tokensCreative tasks
eva-qwen-2.5-72b8K tokensCreative tasks
hermes-3-llama-3.1-70b8K tokensCreative tasks
deepseek-chat32K tokensConversational AI

Embedding

Here are the embedding models supported by Julep:

Model NameEmbedding DimensionsBest For
text-embedding-3-large1024High-quality vectors
voyage-multilingual-21024Cross-language tasks
voyage-31024Advanced embeddings
Alibaba-NLP/gte-large-en-v1.51024Cost-effective solutions
BAAI/bge-m31024Cost-effective solutions
vertex_ai/text-embedding-0041024Google Cloud integration

Though the models mention above support different embedding dimensions, Julep uses fixed 1024 dimensions for all embedding models for now. We plan to support different dimensions in the future.

Supported Parameters

These parameters allow fine-tuning of model behavior. Note that not all parameters are supported by every model.

Best Practices:

  • Start with default values and adjust based on your needs
  • Use temperature (0.0 - 1.0) for most cases
  • Avoid setting multiple penalty parameters simultaneously
  • Test different combinations for optimal results

Setting extreme values for multiple parameters may lead to unexpected behavior or poor quality outputs.

Usage Guidelines

For more information, please refer to the LiteLLM documentation.