Ollama has just been released for linux, which means it’s now dead simple to run large language models on any linux server you choose. I show you how to install and configure it on digitalocean.
00:00 Installation on DigitalOcean
03:30 Running Llama2 on a Server
05:43 Calling a Model Remotely
12:26 Conclusion
#llm #machinelearning
Link:
Support My Work:
Get $200 credit on SignUp to DigitalOcean:
Check out my website:
Follow me on twitter:
Subscribe to my newsletter:
Tip me:
Learn how devs make money from Side Projects:
Gear I use:
14″ Macbook Pro (US) –
14″ Macbook Pro (UK) –
Shure MV7 USB Mic (US) –
Shure MV7 USB Mic (UK) –
As an affiliate I earn on qualifying purchases at no extra cost to you.
[ad_2]
source