NETWORK ADMINISTRATIONSwindows dns serverWindows server

Mistral v3 Released! Did it Pass the Coding Test?

πŸŽ‰ Welcome back, AI enthusiasts! Today, we’re diving deep into the latest version of the Mistral 7B model, version 0.3! With a whopping 32,000 context window, improved tokenizer, and function calling support, this update promises significant advancements. πŸš€ Mistral v3 Released! Did it Pass the Coding Test?

Massed compute: https://bit.ly/mervin-praison
Coupon: MervinPraison (50% Discount)
Connect to Massed Compute after Deploy: https://www.cendio.com/thinlinc/download/

In this video, we’ll:
Compare Mistral 7B v0.3 with Llama 38B across various benchmarks πŸ†š
Test Mistral’s coding abilities with Python challenges 🐍
Evaluate its logical and reasoning skills 🧠
Assess safety features and function calling capabilities πŸ”’

πŸ”§ Setup Guide:
Clone the Repository: git clone https://github.com/oobabooga/text-generation-webui
Navigate to Folder: cd text-generation-webui
Export Hugging Face Token: export HUGGINGFACE_TOKEN=your_token
Start Installation: bash start_linux.sh
Load the Model: Enter the model name and grant access.

πŸ”— Links:
Patreon: https://patreon.com/MervinPraison
Ko-fi: https://ko-fi.com/mervinpraison
Discord: https://discord.gg/nNZu5gGT59
Twitter / X : https://twitter.com/mervinpraison
Sponsor a Video or Do a Demo of Your Product: https://mer.vin/contact/
https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3

πŸ” We’ll be using the unquantized version directly from Hugging Face, and testing its capabilities in various scenarios. Stay tuned as we push this model to its limits!

πŸ‘ If you find this video helpful, don’t forget to like, share, and subscribe for more AI content! Hit the bell icon πŸ”” to stay updated.

πŸ“Œ Timestamps:
0:00 – Introduction to Mistral 7B v0.3
0:11 – Comparison with Llama 38B
0:33 – Setup and Configuration
2:30 – Coding Ability Tests
5:32 – Logical & Reasoning Skills
7:16 – Safety Test
8:00 – Function Calling Demonstration
10:00 – Final Thoughts and Conclusion

source

by Mervin Praison

windows server dns

9 thoughts on β€œMistral v3 Released! Did it Pass the Coding Test?”

  • Great video! liked and shared (already subscribed). But what is the best function calling model right now? Do you have a web or index with the ranking of your models you review? thank you!

  • 🎯 Key Takeaways for quick navigation:

    00:00 πŸŽ‰ Mistral 7B version 0.3 released with 32,000 context window and better tokenizer, supporting function calling.
    00:14 πŸ“Š Llama 38B instruct slightly outperforms Mistral 7B v0.3 in Medical QA accuracy.
    00:24 πŸ” Comparison of Mistral versions 1, 2, and 3 performance, with v0.3 slightly better at coding than v0.2.
    00:38 🀝 Mistral 7B v0.3 and Llama 38B are equally competitive in performance.
    01:06 πŸ’» Testing Mistral 7B v0.3 for coding ability, reasoning skills, and accessing the model on text generation web UI.
    01:33 πŸ› οΈ Setting up the environment and loading the unquantized Mistral 7B v0.3 model.
    03:25 πŸ§ͺ Coding test: Mistral 7B v0.3 performed well on easy challenges but struggled with medium, hard, and expert challenges.
    06:08 🧠 Logical and reasoning test: Mistral 7B v0.3 answered some questions correctly but failed on others.
    07:16 πŸ”’ Safety test: Mistral 7B v0.3 avoided providing information on illegal activities, showing some level of safety.
    08:12 πŸ›°οΈ Function calling test using PraisAI tool: Mistral 7B v0.3 demonstrated function calling ability with Crew AI framework but struggled with Autogen.

    Made with HARPA AI

Comments are closed.