BlackSnack

joined 2 weeks ago
[–] BlackSnack@lemmy.zip 1 points 13 hours ago

Hell ya! I would definitely appreciate some hand holding thru this process! This self hosting stuff is going to take a bit longer and more learning than I anticipated.

-the opening the port process makes sense. It seems like if I have a backend on my rig, I’m going to need to open a port to access that backend from a front end of a phone device. Or possibly even access that same backend on the phone device via a mirror?

-it seems like it would be easier if I could connect to the rig via an android phone instead of an iPhone. My end goal is to use Linux but I’m not ready for that step. Seems like android would be an adequate stepping stone to move to, especially if we have to go thru all this trouble with iPhone. Shall we try on the android instead? If not I’ll follow the directions you put above and report back on Saturday.

[–] BlackSnack@lemmy.zip 1 points 1 day ago (2 children)

Hate to say it but it didn’t work. I listed below the things I double checked. I really appreciate you helping me troubleshoot this, but it seems like I may have bitten off more than I can chew. I choose Ollama because it was supposed to be one of the easier loca AIs to set up. Do you have any recommendations for alternatives? Or do you think I should incorporate a docker or open web ui as some others have said ?

-when I went to the ollama app and entered the http://10.#.#.#:11434 , it didn’t work. Also tried the enchanted app and that didn’t work as well.

-I double checked the rule I made to make sure that was inputted properly. The 10.0.0.0/24 for the local and remote ip addresses.

-the sanity check went well. The ipv4 said closed. The ipv6 said failed.

-I checked the netstat -abn thing and 0.0.0.0:11434 is still listening.

[–] BlackSnack@lemmy.zip 2 points 1 day ago* (last edited 1 day ago) (4 children)

There are 3 lines with the :11434 in them. No brackets or anything like that. -1 has 0.0.0.0 in front -2 has 10.#.#.# in front and has a foreign address that is something other than 0.0.0 -3 is like the 2nd but a slightly different foreign address

The iPhone does have a 10.#.#.# ip number that is slightly different than the PCs.

The subnet mask is 255.255.255.0

Oh yes. I’m on windows 10 as well.


I have taken a pause here while we trouble shoot the subnet mask. We’re getting close!!

[–] BlackSnack@lemmy.zip 2 points 2 days ago (6 children)

Dope! This is exactly what I needed! I would say that this is a very “hand holding” explanation which is perfect because I’m starting with 0% knowledge in this field! And I learned so much already from this post and your comment!

So here’s where I’m at, -A backend is where all the weird c++ language stuff happens to generate a response from an AI. -a front end is a pretty app or webpage that takes that response and make it more digestible to the user. -agreed. I’ve seen in other posts that exposing a port on windows defender firewall is the easiest (and safest?) way to go for specifically what I’m looking for. I don’t think I need to forward a port as that would be for more remote access. -I went to the whatismyipaddress website. The ipv6 was identical to one of the ones I have. The ipv4 was not identical. (But I don’t think that matters moving forward.) -I did the ipconfig in the command prompt terminal to find the info and my ipv4 is 10.blahblahblah.

  • I ran netstat -abn (this is what worked to display the necessary info). I’m able to see 0.0.0.0 before the 11434! I had to go into the settings in the ollama backend app to enable “expose Ollama to the network”.

I’m ready for the next steps!

[–] BlackSnack@lemmy.zip 1 points 2 days ago (2 children)

Bet. Looking into that now. Thanks!

I believe I have 11g of vram, so I should be good to run decent models from what I’ve been told by the other AIs.

[–] BlackSnack@lemmy.zip 0 points 2 days ago (4 children)

Sever is my rig which is running windows. Phone is iPhone.

Exposing the port is something I’ve tried to do in the past with no success! When you say, change the bind address, do I do that in the windows defender firewall in the inbound rules section?

[–] BlackSnack@lemmy.zip 1 points 2 days ago (8 children)

Backend/ front end. I see those a lot but I never got an explanation for it. In my case, the backend would be Ollama on my rig, and the front end would be me using it on my phone, whether that’s with and app or web ui. Is that correct?

I will add kobold to my list of AIs to check out in the future. Thanks!

Ollama has an app (or maybe interface is a better term for it) on windows right that I download models too. Then I can use said app to talk to the models. I believe Reins: Chat for Ollama is the app for iPhone that allows me to use my phone to chat with my models that are on the windows rig.

[–] BlackSnack@lemmy.zip 2 points 2 days ago (10 children)

Yes exactly! I would love to keep it on my network for now. I’ve read that “exposing a port” is something I may have to do in my windows firewall options.

Yes I have Ollama on my windows rig. But im down to try out a different one if you suggest so. TBH, im not sure if librechat has a web ui. I think accessing the LLM on my phone via web browser would be easiest. But there are apps out there like Reins and Enchanted that I could take advantage of.

For right now I just want to do whatever is easiest so I can get a better understanding of what I’m doing wrong.

[–] BlackSnack@lemmy.zip 2 points 2 days ago (12 children)

Oh! Also, I’m using windows on my PC. And my phone is an iPhone.

I’m not using Linux yet, but that is in my todo list for the future! After I get more comfortable with some more basics of self hosting.

[–] BlackSnack@lemmy.zip 1 points 2 days ago

Bet, I’ll try that when I get home tonight. If I don’t have success can I message you directly ?

[–] BlackSnack@lemmy.zip 1 points 2 days ago

Bet. I believe what you mentioned is best for accessing my LLM no matter where I am in the world, correct? If so I will try this one after I try what the other person suggested.

Thank you!

[–] BlackSnack@lemmy.zip 4 points 2 days ago (4 children)

lol I have! They all say the same similar thing but it’s just not working for me.

 

tl-dr

-Can someone give me step by step instructions (ELI5) on how to get access to my LLM's on my rig from my phone?

Jan seems the easiest but I've tried with Ollama, librechat, etc.

.....

I've taken steps to secure my data and now I'm going the selfhosting route. I don't care to become a savant with the technical aspects of this stuff but even the basics are hard to grasp! I've been able to install a LLM provider on my rig (Ollama, Librechat, Jan, all of em) and I can successfully get models running on them. BUT what I would LOVE to do is access the LLM's on my rig from my phone while I'm within proximity. I've read that I can do that via wifi or LAN or something like that but I have had absolutely no luck. Jan seems the easiest because all you have to do is something with an API key but I can't even figure that out.

Any help?

 

EDIT- The issue is having is with “Authentication”. I haven’t made it past that step. Incorrectly said “addresses” on the original post.


Yo yo! Fairly new to making a change towards privacy. My brother gave me a raspberry pi 4 and I want to experiment with that before making a bigger change towards my other electronics. Rn I’m working on using radicale on the pi4 and I’m stuck because I can’t understand the technological language.

I’m trying to follow the tutorial on the radicale website but am getting stuck in the “~~addresses~~” authentication part. I can’t enter in anything Into the initial command prompt I used to create the radicale website. And when I make a new command prompt and enter that in nothing happens. I asked AI and it spits out an answer that isn’t dumbed down enough for me. Lemmy is my last hope before I try Reddit …

Tutorial link for clarity https://radicale.org/v3.html#tutorials

view more: next ›