Ollama on Linux: Easily Install Any LLM on Your Server
Ollama has just been released for linux, which means it’s now dead simple to run large language models on any linux server you choose. I show you how to install and configure it on digitalocean.
00:00 Installation on DigitalOcean
03:30 Running Llama2 on a Server
05:43 Calling a Model Remotely
12:26 Conclusion
#llm #machinelearning
Link:
Support My Work:
Get $200 credit on SignUp to DigitalOcean:
Check out my website:
Follow me on twitter:
Subscribe to my newsletter:
Tip me:
Learn how devs make money from Side Projects:
Gear I use:
14″ Macbook Pro (US) –
14″ Macbook Pro (UK) –
Shure MV7 USB Mic (US) –
Shure MV7 USB Mic (UK) –
As an affiliate I earn on qualifying purchases at no extra cost to you.
[ad_2]
source