Welcome to this month's Fine-Tuned! Today we're introducing the Fine-tuning Index, a comprehensive assessment of 13 popular open-source LLMs and several leading commercial LLMs evaluated across 31 diverse tasks. Read on to learn more!
Spotlight
We're thrilled to release the Fine-tuning Index! This resource shows how fine-tuning open-source LLMs can give a significant performance boost in production environments and ranks top open-source and commercial LLMs based on performance across various tasks.
We designed the Fine-tuning Index with enterprise AI teams in mind, helping you pick the best open-source models for your unique needs. It reports on the performance of 13 popular open-source LLMs across 31 diverse tasks, and even compares them to leading commercial models like GPT-4. Check it out!
Upcoming Events
Fine-tuning an open-source LLM with speculative decoding typically increases inference throughput by more than 2x without sacrificing performance. We’re excited to bring Medusa into the Predibase platform in our next release and invite you to join our upcoming webinar for an early preview and technical Q&A with our CTO Travis Addair and ML Engineer Arnav Garg.
Speed-up LLM Development with Gretel and Predibase
In this workshop, you'll learn how to leverage Gretel and Predibase together to quickly and cost-effectively train LLMs that outperform commercial options. You’ll see how to generate synthetic training data with Gretel Navigator and leverage Predibase to fine-tune an open source LLM using state-of-the-art techniques.
How we accelerated LLM fine-tuning by 15x in 15 days
We’ve taken our experiences fine-tuning 1,000s of LLMs to build a state-of-the-art fine-tuning and serving stack and now we’re sharing those best practices with you. You’ll learn about optimization techniques like Flash attention 2, CUDA kernes, batch size tuning, and more.
In this hands-on virtual workshop, we provide an intro to fine-tuning including use cases, best practices, and techniques for efficient fine-tuning with LoRA. You'll learn how to to use Predibase with $25 in free credits to efficiently fine-tune task-specific models that rival GPT-4 for a series of customer service use cases. Then we show you how to dynamically serve many fine-tuned adapters in real time—all on a single GPU.
How-to Fine-tune & Serve Llama 3 for Automated Customer Support
In this tutorial, we provide a detailed walkthrough of fine-tuning and serving Llama 3 for a customer support use case using Predibase’s new fine-tuning stack. You’ll learn how to easily and efficiently fine-tune and serve open-source LLMs that perform on par with much larger commercial models for task specific use cases.
Try Fine-tuning with up to 10x Faster Training Times on Predibase
In the past, we used Ludwig in Predibase for all fine-tuning tasks to offer a fast and efficient training engine. With this update, we're now using a new fine-tuning system and a special A100 cluster. This means our training engine is even better and can provide users with the quickest fine-tuning speeds available.