I don’t consider myself very technical. I’ve never taken a computer science course and don’t know python. I’ve learned some things like Linux, the command line, docker and networking/pfSense because I value my privacy. My point is that anyone can do this, even if you aren’t technical.

I tried both LM Studio and Ollama. I prefer Ollama. Then you download models and use them to have your own private, personal GPT. I access it both on my local machine through the command line but I also installed Open WebUI in a docker container so I can access it on any device on my local network (I don’t expose services to the internet).

Having a private ai/gpt is pretty cool. You can download and test new models. And it is private. Yes, there are ethical concerns about how the model got the training. I’m not minimizing those concerns. But if you want your own AI/GPT assistant, give it a try. I set it up in a couple of hours, and as I said… I’m not even that technical.

176 points

“learned some things like Linux, command line, docker, and networking/pfsense” “I don’t consider myself technical”

Don’t sell yourself short, I work in IT and have colleagues on our helpdesk who would struggle endlessly with those concepts.

I hereby dub you a tech person, like it or not, those skills can and do pay the bills.

permalink
report
reply
61 points

Now that you’ve dubbed OP a tech person…

Hey OP, can you help me fix my printer? It’s only printing “RED RUM RED RUM” for some reason.

permalink
report
parent
reply
21 points

Have you tried giving it red rum?

Oh, and make sure you hold it out with the insides of your arms exposed, it’ll feel less threatening that way.

permalink
report
parent
reply
2 points

Have you replaced the blood cartridge?

permalink
report
parent
reply
43 points

It is done.

permalink
report
parent
reply
40 points

This made me smile. Thank you. The grass is always greener and I sometimes daydream of working in IT instead of healthcare. Maybe someday.

permalink
report
parent
reply
13 points

Nah dont.

permalink
report
parent
reply
8 points

Healthcare is pretty rough, I’d be willing to bet that the grass actually is greener in this case.

permalink
report
parent
reply
5 points

hahahaha best advice ever.

permalink
report
parent
reply
4 points

Join us. We have cookies (well at least until the end of our sessions)!

permalink
report
parent
reply
11 points

Thank you for this. I consider myself technical and those words felt like a punch in the gut.

permalink
report
parent
reply
12 points

I’m sorry if I offended. I can’t code or understand existing code and have always felt that technical people code. I guess I should expand my definition. Again, sorry that my words felt like a punch in the gut… wasn’t my intention at all.

permalink
report
parent
reply
11 points

It depends heavily on what you do and what you’re comparing yourself against. I’ve been making a living with IT for nearly 20 years and I still don’t consider myself to be an expert on anything, but it’s a really wide field and what I’ve learned that the things I consider ‘easy’ or ‘simple’ (mostly with linux servers) are surprisingly difficult for people who’d (for example) wipe the floor with me if we competed on planning and setting up an server infrastructure or build enterprise networks.

And of course I’ve also met the other end of spectrum. People who claim to be ‘experts’ or ‘senior techs’ at something are so incompetent on their tasks or their field of knowledge is so ridiculously narrow that I wouldn’t trust them with anything above first tier helpdesk if even that. And the sad part is that those ‘experts’ often make way more money than me because they happened to score a job on some big IT company and their hours are billed accordingly.

And then there’s the whole other can of worms on a forums like this where ‘technical people’ range from someone who can install a operating system by following instructions to the guys who write assembly code to some obscure old hardware just for the fun of it.

permalink
report
parent
reply
10 points

This gave me confidence as well, thank you 😆

permalink
report
parent
reply
8 points

I was just talking to a member of my devops team and I was talking about this exact thing and they said “I didn’t know you could attach a GPU to a container”. So, yup, just stay on top of this stuff at home and you’ll do fine

permalink
report
parent
reply
2 points

Who uses GPUs for AI anyway. They cost more than a car sometimes

permalink
report
parent
reply
1 point

Can confirm, the GPU in my laptop costs more than all but my newest car.

permalink
report
parent
reply
7 points

With how low the average person is with tech skills, it’s very easy to be top 10%.

permalink
report
parent
reply
49 points

people need to take a step back and realize we have the capability to trap quasi-omnipotent quasi-demons in our personal computers

yeah they lie a lot and rarely do what you want them to, but that’s just what demons do

And it’s all powered by some dark crystals created with light magic that slowly poison the planet

that’s some arcane bullshit

permalink
report
reply
3 points

How long can something like that really last, though? I wish we had a better idea of the timeline, before the quasi-demons start freelancing lol

permalink
report
parent
reply
28 points

Uncensored models are so much better, too. chatGPT is like one of those plastic children’s toy hammers vs real models are titanium hammers

permalink
report
reply
6 points

Together.ai has a number of uncensored models too. I’ve found that those are so cheap that it’s not worth trying to self just models unless you really need more privacy.

permalink
report
parent
reply
15 points

Yeah, I like it too. My only issue is ollama’s lack of intel support. I have been looking at issue 1590 on their GitHub. For now I have a 1050ti in a cardboard box PC with other hardware being 10+ years old and a mixed set of RAM totalling 12G. It also has a 100Mbit nic, so I can’t take advantage of full internet speed when downloading models. The worst part is they can support intel, but haven’t merged the solution because of an issue with the windows intel drivers. Linux is fine but I can 't have it. I wasn’t planning to rant, but I already typed it so… enjoy?

permalink
report
reply
11 points
*

Yeah, I have an NVDIA GPU and it is magic. The best part is when you are using Ollama, open a second terminal window and enter the command, watch -n 0.5 nvidia-smi and you can see your GPU usage go up and down in real-time as you ask the GPT questions. Pretty cool.

Hopefully they get the ARC folks up and running soon.

permalink
report
parent
reply
3 points

You can also achieve this with nvtop and have a pretty TUI (terminal UI).

permalink
report
parent
reply
3 points
*

There is ipex-llm from Intel which you can use with your intel IGPU/GPU/CPU for llms which also supports ollama.

permalink
report
parent
reply
10 points

Have you found much practical use for small models yet? I love the idea that even the 1.1B tinyllama model can run on my phone, but haven’t found much real world use for it yet. Llama3 8b feels better, but not much better for even emails as it’s a bit dumb

permalink
report
reply
6 points

I use my phone all the time, but I just use a wireguard VPN to tunnel into my home container of Open WebUI. Then I can interact with my desktop machine using a NVIDIA gpu. I’m currently testing mistral-nemo. It’s pretty great but it gets a bit verbose sometimes.

permalink
report
parent
reply
10 points

I am also using open webui. Most LLMs are too verbose for me, so I created a model in open-webui with system prompt “Do not repeat the questions. Avoid giving lists as answers. Do not summarize the answer at the end. If asked a follow-up question, respond with only new information, do not repeat previously stated information.” and named it No Nonsense.

permalink
report
parent
reply
3 points

That’s really smart. I just found out about fabric yesterday and it is helping me with things like what you stated. Prompt engineering is a huge thing.

permalink
report
parent
reply
2 points

for some reason chatgpt responds well to “no yapping”

permalink
report
parent
reply
1 point

Imo it’s worthwhile to just run the biggest model available and rent expensive GPU time. It still amounts to very little overall and you get much better results. Project dependent of course

permalink
report
parent
reply

Selfhosted

!selfhosted@lemmy.world

Create post

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don’t control.

Rules:

  1. Be civil: we’re here to support and learn from one another. Insults won’t be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it’s not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don’t duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

Community stats

  • 3.4K

    Monthly active users

  • 1.6K

    Posts

  • 14K

    Comments