NETWORK ADMINISTRATIONSwindows dns serverWindows server

Mistral v3 Released! Did it Pass the Coding Test?

🎉 Welcome back, AI enthusiasts! Today, we’re diving deep into the latest version of the Mistral 7B model, version 0.3! With a whopping 32,000 context window, improved tokenizer, and function calling support, this update promises significant advancements. 🚀 Mistral v3 Released! Did it Pass the Coding Test?

Massed compute: https://bit.ly/mervin-praison
Coupon: MervinPraison (50% Discount)
Connect to Massed Compute after Deploy: https://www.cendio.com/thinlinc/download/

In this video, we’ll:
Compare Mistral 7B v0.3 with Llama 38B across various benchmarks 🆚
Test Mistral’s coding abilities with Python challenges 🐍
Evaluate its logical and reasoning skills 🧠
Assess safety features and function calling capabilities 🔒

🔧 Setup Guide:
Clone the Repository: git clone https://github.com/oobabooga/text-generation-webui
Navigate to Folder: cd text-generation-webui
Export Hugging Face Token: export HUGGINGFACE_TOKEN=your_token
Start Installation: bash start_linux.sh
Load the Model: Enter the model name and grant access.

🔗 Links:
Patreon: https://patreon.com/MervinPraison
Ko-fi: https://ko-fi.com/mervinpraison
Discord: https://discord.gg/nNZu5gGT59
Twitter / X : https://twitter.com/mervinpraison
Sponsor a Video or Do a Demo of Your Product: https://mer.vin/contact/
https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3

🔍 We’ll be using the unquantized version directly from Hugging Face, and testing its capabilities in various scenarios. Stay tuned as we push this model to its limits!

👍 If you find this video helpful, don’t forget to like, share, and subscribe for more AI content! Hit the bell icon 🔔 to stay updated.

📌 Timestamps:
0:00 – Introduction to Mistral 7B v0.3
0:11 – Comparison with Llama 38B
0:33 – Setup and Configuration
2:30 – Coding Ability Tests
5:32 – Logical & Reasoning Skills
7:16 – Safety Test
8:00 – Function Calling Demonstration
10:00 – Final Thoughts and Conclusion

source

by Mervin Praison

windows server dns

9 thoughts on “Mistral v3 Released! Did it Pass the Coding Test?

  • Great video! liked and shared (already subscribed). But what is the best function calling model right now? Do you have a web or index with the ranking of your models you review? thank you!

  • 🎯 Key Takeaways for quick navigation:

    00:00 🎉 Mistral 7B version 0.3 released with 32,000 context window and better tokenizer, supporting function calling.
    00:14 📊 Llama 38B instruct slightly outperforms Mistral 7B v0.3 in Medical QA accuracy.
    00:24 🔍 Comparison of Mistral versions 1, 2, and 3 performance, with v0.3 slightly better at coding than v0.2.
    00:38 🤝 Mistral 7B v0.3 and Llama 38B are equally competitive in performance.
    01:06 💻 Testing Mistral 7B v0.3 for coding ability, reasoning skills, and accessing the model on text generation web UI.
    01:33 🛠️ Setting up the environment and loading the unquantized Mistral 7B v0.3 model.
    03:25 🧪 Coding test: Mistral 7B v0.3 performed well on easy challenges but struggled with medium, hard, and expert challenges.
    06:08 🧠 Logical and reasoning test: Mistral 7B v0.3 answered some questions correctly but failed on others.
    07:16 🔒 Safety test: Mistral 7B v0.3 avoided providing information on illegal activities, showing some level of safety.
    08:12 🛰️ Function calling test using PraisAI tool: Mistral 7B v0.3 demonstrated function calling ability with Crew AI framework but struggled with Autogen.

    Made with HARPA AI

Comments are closed.