Getting Started with OLLAMA – the docker of ai!!!
chris explores how ollama could be the docker of AI. in this video he gives a tutorial on how to get started with ollama and run models locally such as mistral-7b and llama-2-7b. he looks at how ollama operates and how it works very similarly to docker including the concept of the model library. chris also shows how you can create customized models and how to interact with the built-i fastapi server as well as use the javascript ollama library to interact with the models using node.js and bun. at the end of this tutorial you’ll have a great understanding of ollama and it’s importance in AI Engineering
by Chris Hay
linux web server
Great video, really enjoyed this! Thanks Chris