Comparing Quantizations of the Same Model - Ollama Course
Matt Williams Matt Williams
31.8K subscribers
4,198 views
277

 Published On Aug 20, 2024

Welcome back to the Ollama course! In this lesson, we dive into the fascinating world of AI model quantization. Using variations of the llama3.1 model, we explore how different quantization levels affect performance and output quality.

Through this video, you'll gain a deeper understanding of how to choose the right quantization for how you use AI models, ensuring you get the best performance and results for your specific needs. Don't forget to subscribe for more lessons in this free Ollama course!
Thanks for watching!


You can find the code for this video at https://github.com/technovangelist/vi...

Be sure to sign up to my monthly newsletter at https://technovangelist.substack.com/...

I have a Patreon at   / technovangelist  

You can find the Technovangelist discord at:   / discord  
The Ollama discord is at   / discord  

(they have a pretty url because they are paying at least $100 per month for Discord. You help get more viewers to this channel and I can afford that too.)

00:00 - Start with an example
00:24 - Introduction
00:56 - Lots of claims on the Discord
01:26 - Intro to the app
01:57 - Where to find the code
02:20 - Grab a few quantizations
02:57 - You should regularly pull the models again
03:30 - Back to the Black Hole answers
04:39 - The classic logic problem
05:35 - How about function calling
08:31 - How about for prompts with more reasoning
09:01 - Are those questions stupid?
09:30 - Which quant to use?

show more

Share/Embed