Proxmox

Run Mistral, Llama2 and Others Privately At Home with Ollama AI – EASY!



Self-hosting Ollama at home gives you privacy whilst using advanced AI tools. In this video I provide a quick tutorial on how to set this up via the CLI and Docker with a web GUI.

Ollama:

Video Instructions:

Recommended Hardware:

Discord:
Twitter:
Reddit:
GitHub:

00:00 – Overview of Ollama and LLMs
01:38 – Creating a VM
02:52 – Installation – CLI
05:50 – Installation – Docker
11:55 – Outro

[ad_2]

source

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button