Make sure to sign-up for our monthly newsletter to receive all the latest AI and LLM news, tutorials and product updates.
We hope your 2024 is off to a great start! 2023 was certainly transformative for AI but we think 2024 will see shifts that make AI even more accessible, reliable, and secure for practitioners. To that end, we’re excited to release our AI and LLM Predictions for 2024, and to announce our first webinar of 2024! Read on for more details.
Featured Event
Fine-Tuning Zephyr-7B to Analyze Customer Support Call Logs
Join us on February 1st at 10:00 am PT to learn how you can leverage open source LLMs to automate one of the most time consuming tasks of customer support: classifying customer issues. You will learn how to efficiently and cost-effectively fine-tune an open source LLM with just a few lines of code at a fraction of the cost of using a commercial LLM and how to easily implement efficient fine-tuning techniques like LoRA and quantization.
Recent Events + Podcasts
How Fine-tuning Open Source LLMs Solves GenAI Productionization - Piero Molino | Stanford MLSys #94
LoRA Bake-off: Comparing Fine-Tuned Open-source LLMs that Rival GPT-4
Featured Blog Post
LoRA Land: Fine-Tuned Open-Source LLMs that Outperform GPT-4
LoRA Land is a collection of 25 fine-tuned Mistral-7b models that consistently outperform base models by 70% and GPT-4 by 4-15%, depending on the task. LoRA Land’s 25 task-specialized large language models (LLMs) were all fine-tuned with Predibase for less than $8.00 each on average and are all served from a single A100 GPU using LoRAX. Learn more!
Customer Spotlight
Deep Learning on Snowflake: How Paradigm Built a Personal Trading Experience with Predibase
Paradigm, one of the world’s largest institutional liquidity networks for cryptocurrencies, was able to use Predibase to build a deep learning-based recommendation system for their traders on top of their existing Snowflake Data Cloud with just a few lines of YAML. Building production models on top of Snowflake data–a task that used to take months–now takes minutes.
From the Community
Fine-tuning LLMs for cost effective GenAI inference at scale
Predibase joins the AI Alliance
Fine Tune mistral-7b-instruct on Predibase with Your Own Data and LoRAX
When to Fine-Tuning Large Language Models?
Open Source Updates
How to Efficiently Fine-Tune Gemma-7B with Open-Source Ludwig
Google recently released Gemma, a state-of-the-art LLM, licensed free of charge for research and commercial use. In this short tutorial, we showed you how to easily, reliably, and efficiently fine-tune Gemma-7B-Instruct on readily available hardware using open-source Ludwig. Try it out and share your results with our Ludwig community on Slack.
Ludwig v0.9.2 & v0.9.3: These releases introduces a few bug fixes and several new enhancements:
- Add support for the official microsoft/phi-2 model
- Ensure correct padding token for Phi and Pythia models
- Cast
LLMEncoder
output totorch.float32
, freeze final layer at init - Enable IA3 adapters
- Add batch size tuning for LLMs
- Per-step token utilization to tensorboard and progress tracker
- Default LoRA target modules for #Mixtral and #Mixtral-instruct
- Support for exporting models to Carton (thank you Vivek Panyam!)
LoRAX v0.6: The latest release adds support for multi-turn chat conversations with dynamic LoRA adapter loading. Just replace the "model" parameter with any HF LoRA and you're set. Chat templates can come from the base model, or even LoRA itself if it uses its own custom template. All of this happens dynamically per request.
Featured Product Update
We’re excited launch our new prompting experience in the Predibase UI which allows you to easily prompt serverless endpoints and your fine-tuned adapters without needing to deploy them first. This lets teams test their fine-tuned models and compare model iterations all from the UI, enabling much faster test and review cycles.