Depending on how much ram you have, choose wisely.
8b > 8~16GB of ram.
70b > 64GB or more
Is it all CPU based or does it offload to the GPU? There a couple older workstations I could get my hands on with still decent CPUs that go up to 128 Gb of RAM, but sourcing a decent GPU for them would be expensive.
It’s both (detected at the installation).
I have two instances running at the same time on one computer.
First is running natively on my GPU, the second on a VM with 8 shared cores.
How do I know which was installed if I simply performed the "ollama run dolphin-llama3" command and it fetched the Dolphin model?
Am I even using the right vernacular? I need to skill up on all this, LOL.
The 8b is selected by default.
You can check by running
ollama run dolphin-llama3:8b
It will either download the model or run it (if you already have it installed)
(post is archived)