I’ve been playing around with Ollama in a VM on my machine and it is really useful.

To get started I would start by making sure you have capable hardware. You will need recent hardware so that old computer you have laying around may not be enough. I created a VM on my laptop with KVM and gave it 8gb of ram and 12 cores.

Next, read the readme. You can find the Readme at the github repo

https://github.com/ollama/ollama

Once you run the install script you will need to download models. I would download Llama2, Mistral and LLava. As an example you can pull down llama2 with ollama pull llama2

Ollama models are available in the online repo. You can see all of them here: https://ollama.com/library

Once they are downloaded you need to setup openwebui. First, install docker. I am going to assume you already know how to do that. Once docker is installed pull and deploy open web UI with this command. Notice its a little different than the command in the open web UI docs. docker run -d --net=host -e OLLAMA_BASE_URL="http://localhost:11434 -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main

Notice that the networking is shared with the host. This is needed for the connection. I also am setting the environment variable in order to point open web UI to ollama.

Once that’s done open up the host IP on port 8080 and create an account. Once that’s done you should be all set.

There’s also llamafile, super simple: download and run it.

Possibly linux
creator
link
fedilink
English
16M

Not as cool or flexiable though

Iirc it can run anything llama.cpp can because it just uses that under the hood.

Possibly linux
creator
link
fedilink
English
26M

Except you can’t control it as easily. I like the UI and toolset of open web UI

Ok. I haven’t tried it so I’ll take your word for it. I’m just offering an easier alternative since the topic was “getting started”

@cron@feddit.de
link
fedilink
English
36M

Thanks for sharing this. Its a shame that most AI tech is hidden behind steep price tags and cloud subscriptions, while even midrange PCs can run interesting AI models.

@PerogiBoi@lemmy.ca
link
fedilink
English
36M

Check out LM Studio. Totally free and very nice GUI. You can do some very cool things with it. That and Kobold AI. Very simple self hosting no need for configuration either. Download and open essentially.

Scrubbles
link
fedilink
English
76M

Personally I’ve really enjoyed text-generation-webui. It made it really easy to ramp up and learn. Very cool stuff you got though, I’ll probably be looking at a comparison between them!

@RedNight@lemmy.ml
link
fedilink
English
36M

Ollama has been great for self-hosting, but also checkout vLLM as its the new shiny self-hosting toy

mozz
link
fedilink
56M

Does it work out okay with 12 cores purely on CPU? About how fast is the interaction?

I played around a little with Ollama and gpt4all but it seemed to me like it wasn’t fast enough to be useful on pure CPU, but if I could just throw cores at it then I might revisit the issue.

@goosegooseboat@lemmy.world
link
fedilink
English
2
edit-2
6M

For the life of me can’t remember the scores I was getting on gpt4all. But given that you tried it I’m guessing you’ll most likely take a liking to LM studio or perhaps jan.ai. both GUI tools. If the lack open source bothers you go for Jan.ai, if not then go LM studio. LM studio in particular allows for full and partial GPU offloading. So if you have a semi capable but not quite enough vram on it you can load part of the model on the GPU to speed up inference. As a side note pure CPU on my old ryzen 1600 I was looking at 6/it. Which isn’t all that much but glass half full its still faster than the average typing speed and takes the load off of having to think about how to creatively word things

Possibly linux
creator
link
fedilink
English
36M

It wasn’t usable a few months ago. However, when I setup ollama it was “fast” and it works ok. It takes anywhere from instant to 5min for responses. LLava seems to take the longest which makes sense. For llama2 it is fairly fast unless you ask it for obscure information.

@thantik@lemmy.world
link
fedilink
English
36M

The biggest thing that I want to learn is how to either A: add “tools” for the AI to run, or B: “fine-tune” the model by feeding it data that’s relevant to me.

Possibly linux
creator
link
fedilink
English
26M

You can teach it things and upload documents for it to process

@thantik@lemmy.world
link
fedilink
English
26M

Yeah, I couldn’t find that on ollama; but I did find it in text-generation-webui - which is a little more complicated, but for me, I think it might help springboard me into understanding a few more things.

Possibly linux
creator
link
fedilink
English
26M

Ollama is just a the backend. You need open web UI or a similar application to use it

@Willdrick@lemmy.world
link
fedilink
English
26M

Check AnythingLLM out, its just an appimage

Possibly linux
creator
link
fedilink
English
16M

Not as maintainable long term and it doesn’t have user management

Possibly linux
creator
link
fedilink
English
1
edit-2
6M

So why is it better than OpenwebUI? It seems like each has there own use case.

I’ll give it a try just for fun but it doesn’t seem to be better as far as I can tell

Create a post

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don’t control.

Rules:

  1. Be civil: we’re here to support and learn from one another. Insults won’t be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it’s not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don’t duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

  • 1 user online
  • 279 users / day
  • 589 users / week
  • 1.34K users / month
  • 4.55K users / 6 months
  • 1 subscriber
  • 3.47K Posts
  • 69.3K Comments
  • Modlog