100+ Open-source LLMs Ready to Use
The largest selection of models for fine-tuning and serving
Predibase offers the largest selection of open-source models for fine-tuning and serving and is also the exclusive fine-tuning partner for Upstage's Solar LLM. Check out our current list of supported models and guidance on how to bring your own models in the table below.
Name | Description | Context Window | Always On |
---|---|---|---|
Llama 3.1 8B Serverless | Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks. | 63,999 | |
Llama 3.1 8B Instruct Serverless | Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks. | 63,999 | |
Llama 3 8B Serverless | Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. This is the base 8B parameter model. | 8,192 | |
Llama 3 8B Instruct Serverless | Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. This is the instruction-tuned 8B parameter model. | 8,192 | |
Llama 3 70B Serverless | Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. This is the base 70B parameter model. | 8,192 | |
Llama 3 70B Instruct Serverless | Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. This is the instruction tuned 70B parameter model. | 8,192 | |
Llama 2 7B Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the base 7B model. | 4,096 | |
Llama 2 7B Chat Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the instruction tuned 7B model. | 4,096 | |
Llama 2 13B Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the base 13B model. | 4,096 | |
Llama 2 13B Chat Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the instruction-tuned 13B model. | 4,096 | |
Llama 2 70B Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the base 70B model. | 4,096 | |
Llama 2 70B Chat Serverless | Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the instruction-tuned 70B model. | 4,096 | |
CodeLlama 7B Serverless | Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the 7B base version. | 16,000 | |
CodeLlama 7B Instruct Serverless | Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the 7B instruct-tuned version. | 16,000 | |
CodeLlama 13B Instruct Serverless | Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the instruction-tuned 13B model. | 4,096 | |
CodeLlama 70B Instruct Serverless | Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the instruction-tuned 70B model. | 4,096 | |
Mistral 7B Serverless | The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. | 32,768 | |
Mistral 7B Instruct Serverless | The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets. | 32,768 | |
Mistral 7B Instruct v02 Serverless | The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2. | 32,768 | |
Mistral 7B Instruct v03 Serverless | The Mistral-7B-Instruct-v0.3 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.3. | 32,768 | |
Mixtral 8x7B v01 Serverless | The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. | 32,768 | |
Mixtral 8x7B Instruct v01 Serverless | The Mixtral-8x7B-instruct-v0.1 Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. | 32,768 | |
solar-1-mini-chat-240612 Serverless | Upstage's Solar model is a family of large language models (LLMs) known for its compact yet powerful performance compared to similar sized models below 13B parameter size. The Solar Chat model is optimized for multi-turn chat purposes to effectively handle extended conversations, making it a compelling choice for interactive applications. | 32,768 | |
solar-pro-preview-instruct Serverless | Solar Pro Preview is an advanced LLM with 22 billion parameters designed to fit into a single GPU. Solar Pro Preview shows superior performance compared to LLMs with less than 30 billion parameters and delivers performance comparable to models over three times its size, such as Llama 3.1 with 70 billion parameters. | 4,096 | |
Zephyr 7B Beta Serverless | Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). | 32,768 | |
Phi 2 Serverless | Phi-2 is a Transformer with 2.7 billion parameters. It was trained using the same data sources as Phi-1.5, augmented with a new data source that consists of various NLP synthetic texts and filtered websites (for safety and educational value). | 2,048 | |
Phi 3 4k Instruct Serverless | The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets. This is the 4k context window instruct version of the model. | 4,000 | |
phi-3-5-mini-instruct Serverless | Phi-3.5-mini is a lightweight, state-of-the-art open model built upon datasets used for Phi-3 - synthetic data and filtered publicly available websites - with a focus on very high-quality, reasoning dense data. | 128,000 | |
Gemma 2B Serverless | This is the 2B base version of the Gemma model. | 8,192 | |
Gemma 2B Instruct Serverless | This is the 2B instruction-tuned version of the Gemma model. | 8,192 | |
Gemma 7B Serverless | This is the 7B base version of the Gemma model | 8,192 | |
Gemma 7B Instruct Serverless | This is the 7B instruction-tuned version of the Gemma model. | 8,192 | |
Gemma 2 9B Serverless | Google's latest iteration of open LLMs, based on Google Deepmind Gemini | 8,191 | |
Gemma 2 9B Instruct Serverless | Google's latest iteration of open LLMs, based on Google Deepmind Gemini | 8,191 | |
Gemma 2 27B Serverless | Google's latest iteration of open LLMs, based on Google Deepmind Gemini | 8,191 | |
Gemma 2 27B Instruct Serverless | Google's latest iteration of open LLMs, based on Google Deepmind Gemini | 8,191 | |
Qwen2 1.5B Instruct Serverless | Qwen2 is a language model series including decoder language models of different model sizes, developed by Qwen team, Alibaba Cloud | 65,536 | |
Qwen2 7B Serverless | Qwen2 is a language model series including decoder language models of different model sizes, developed by Qwen team, Alibaba Cloud | 131,072 | |
Qwen2 7B Instruct Serverless | Qwen2 is a language model series including decoder language models of different model sizes, developed by Qwen team, Alibaba Cloud | 131,072 | |
Qwen2 72B Serverless | Qwen2 is a language model series including decoder language models of different model sizes, developed by Qwen team, Alibaba Cloud | 131,072 | |
Qwen2 72B Instruct Serverless | Qwen2 is a language model series including decoder language models of different model sizes, developed by Qwen team, Alibaba Cloud | 131,072 |
Bring Your Own Model
Predibase is the only platform that allows you to take nearly any standard base or fine-tuned LLM from HuggingFace and serve it as a private serverless deployment on state-of-the-art managed infra.