I've finally powered on a 15 year old machine to run a bot I've been writing. The thing is slow as dirt and stuck behind a flakey power line network, but it's working. I got to write my first systemd service definition, which is kind of cool.
Selfhosted
A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.
Rules:
-
Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.
-
No spam posting.
-
Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.
-
Don't duplicate the full text of your blog or github here. Just post the link for folks to click.
-
Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).
-
No trolling.
Resources:
- selfh.st Newsletter and index of selfhosted software and apps
- awesome-selfhosted software
- awesome-sysadmin resources
- Self-Hosted Podcast from Jupiter Broadcasting
Any issues on the community? Report it using the report flag.
Questions? DM the mods!
The computer I'm using currently, I set the BIOS in 2012. WHen I built it, I stuffed every last piece of cutting edge tech of the time into it. Dual CPU, SLI, started with 64gb ram then later on maxed the board out at 128gb. It's still a workhorse tho. It's one of the three I use all the time for music production, selfhosting etc.
My machine is not a workhorse. I got it second hand. It has around 8gb of RAM, and an 80gb HDD I found in a laptop.
But it's enough to work as a testbed, so it's fine with me.
Finally switched from plex to jellyfin, seems to be ok so far. Needed to make some small scripts for metadata management but it's running smoothly. Finally decided I'm hosting enough software with user accounts that I've made an authentik instance for SSO with each (ofc jellyfin first)
Ann reason you choose authenik? There are a nmber of options and I'm not sure why to choose one over the other.
I'm not the person you're replying to, but Authentik:
- Has a UI for configuring it, including adding users.
- Supports LDAP if you need it. Authelia needs a separate LDAP server.
- Supports practically every two factor auth protocol you'd need: OIDC (OpenID Connect), OAuth2, SCIM, SAML, RADIUS, LDAP, and proxying for apps that don't support any of them (which is getting rarer).
- Supports permissions and permission groups, i.e. only allow certain users to access particular apps.
- Can be used as the source of truth for Google Workspace and Microsoft Entra. Maybe not as relevant for home use.
I haven't tried Keycloak but I hear it's pretty good, albeit a heavier app to deploy.
I have tried Authelia, and it's much less powerful than Authentik. Authelia requires you to manually modify config files rather than using a web UI. It also only supports OIDC (which is in beta) and proxying. Proxying is not recommended and has several issues since it's not "true" single sign-on.
I did no research whatsoever and picked the one I'd seen the name of more often. I figured if it didn't work for me I'd try something else, same as when plex wasn't working for me so I switched to jellyfin. I have no idea how it compares to the other options but it feels pretty solid so far
Hey, we're also thinking about setting up authentik. Could you answer the following, where I haven't found answers to yet: does introducing SSO impede logging into Jellyfin on a TV / phone app at all?
no, works fine. there’s an LDAP plugin for jellyfin so you can use the jellyfin internal login page and the server will verify the login against authentik. took some setting up though.
Found out that docker volumes are important after restarting my server 🙃
Am I mistaken that docker creates temporary volumes with a nondescript name and you can potentially dig up the volumes that were being used in /var/lib/docker/volumes
?
Oh, I've just been tinkering around with LangFlow specifically as a news aggregator.
The flow: https://i.imgur.com/5HqznQm.png
Then asking AI to go get me some news: https://i.imgur.com/ltZPBwC.png
Still needs a little tinkering and as the final step, to send said news stories to my Telegram. I really have a blast with automation platforms like N8N, Flowise, Gotify, DopplerTask, & Kestra.
Afterwards, I smoked a small bowl and worked on a couple songs I have in the works.
HBU?
More incus:
- mounting persistent storage into containers (cheating by exporting NFS from my proxmox zfs into the incus host.
- wrote a pruning backup script for containers, runs daily
- passed through hardware (quicksync) into jellyfin container (it works!)
- launched an OCI container (docker home assistant) natively in incus (this is a game-changer!)
Next:
- build 2nd incus node
- move all containers from proxmox to incus
- decom proxmox
- setup Debian with NFS export
Recently been working on setting up forgejo to migrate away from GitHub. My open source stuff I’ve actually put onto codeberg and I’ve set up a handful of pull mirrors on my local instance for redundancy. This weekend I’ve been testing out woodpecker-ci for automating pushing files to s3 for some static websites for repos on codeberg as well as my forgejo instance. Today will tell if that is successful!
I set up my own Lemmy server, mastodon, and matrix. Finally making the move off centralized social media and communication platforms
Do you just do this for your own personal use, a few friends or just anyone from the internet?I'm just curious what the point is and how much effort is involved in connecting with other instances.
Finished my migration from Plex to Jellyfin
A catalog for organizing various Roms you have. It can pull metadata from a number of sources and properly add all the details, cover art, and platform information to each game. It’s smart enough to auto-generate collections based on game series, and embed YouTube videos for gameplay of each one without even any configuration.
The best part? It has Ruffle and EmulatorJS built in so you can play any games supported by EmulatorJS in your browser. I tested games up to N64 and they all ran smooth as butter right in the browser with gamepad configurations built in. They even support local multiplayer.
This week I finally managed to route torrent traffic through a VPS that was sitting around gathering dust. I am behind CGNAT so was taking me 6 weeks to do the kind of traffic I do in a day now. I couldn't be more chuffed.
Scrubbing a little demo project I made featuring a web app behind oauth2-proxy leveraging keycloak as local idp with social login. It also uses a devcontainer config for development. The demo app uses the Litestar framework (fka starlite, in Python) because I was interested, but it's hardly the focus. Still gotta put caddy in front of it all for easy SSL. Oh, and clean up all the default secrets I've strewn about with appropriate secret management.
All of it is via rootless podman and declarative configuration.
Think I might have to create my own Litestar RBAC plugin that leverages the oauth headers provided by the proxy.
It has been a minute since I worked daily in this space, so it has been good to dust off the cobwebs.
I'm switching my immich instance to an SSD one and switching my VPN from zerotier to tailscale.
Hopefully that means my Immich will be a little more reactive.
If at all possible see if you can do wireguard yourself. Tailscale is basically inserting a third party company for no reason as its just wireguard with their servers involved. For example if you can run opnsense its easy to get running via the GUI. Very rewarding!
I've been testing out immutable distros, in this case openSUSE Aeon (laptop) and openSUSE MicroOS (server).
I set up Forgejo and runners are working, all in podman. I'm about to take the plunge and convert everything on my NAS to podman, which is in preparation for installing MicroOS on it (upgrade from Leap).
I also installed MicroOS on a VPS, which was a pain because my VPS provider doesn't have images for it, and I'd have to go through support to get it added. Instead, I found a workaround, which is pretty amazing that it works:
- Install Alpine Linux (in my case I needed to provision something else first and mount an ISO to install Alpine, which was annoying)
- Download MicroOS image on VPS (not ISO, qcow image)
- Write image to the disk, overwriting the current OS (qemu-img command IIRC)
- Reboot (first boot takes longer since it's expanding the disk and whatnot)
The nice thing is that cloud-init works, so my keys set up in step 1 still work with the new OS. It's not the most convenient way to set things up, but it's about the same amount of time as asking them for an ISO.
Anyway, now it's the relatively time consuming task of moving everything from my other VPS over, but I'll do it properly this time with podman containers. I had an ulterior motive here as well, I'm moving from x86 to ARM, which reduces cost somewhat and it can also function as a test bed of sorts for ARM versions of things I'm working on.
So far I'm liking it, especially since it forces me to use containers for everything. We'll see in a month or two how I like maintaining it. It's supposed to be super low effort, since updates are installed in the background and applied on reboot.
I've just moved and I'm setting up my machines. NIC died in my DIY router just before the move so I'm upgrading to 2.5/10 Gbps at the same time.
Email... My wife really wants to further de-google, this means moving custom domains off gsute.
Do I move to proton/tuta or go back to self hosting email again like I did for years until about 2010?
If I self host, do I do it at home or on the server that runs my lemmy instance?
I self-host my email using Mailcow, and use a VPS for it. I don't trust my home server to be reliable enough, and the VPS providers have nicer equipment (modern AMD EPYC CPUs, enterprise SSDs, datacenter-grade 10Gbps or 40Gbps connections, etc). I use a separate VPS just for my emails - it's the one thing I want to ensure is secure, so I didn't want any other random software (that could potentially have security issues) running on it..
I also use an outbound SMTP relay to avoid having to deal with IP reputation. Very easy to configure this in Mailcow. SMTP2Go has a free plan for sending <1000 emails per month.
It kind of amazes me that, in this day and age, email has turned out to be the lynchpin of security. Email as a 2FA endpoint. Email password reset systems. If email is compromised, everything else falls. They used to tell us not to put anything in email that you wouldn't put on a postcard...how did this happen?
Got my jetKVM in the mail yesterday. Really sleek build and software. Liking it a lot so far.
Migrated my network to a router running openwrt this past week as well. Having issues with avahi-daemon crash looping, so I haven't been able to get mdns working in between networks 🤷
I tried to update my lemmy instance and it all went so horribly wrong. DB never came up, errors everywhere, searching implied I updated to a dev branch sometime in the past (not a dev, don't think I did) and it'll be console and DB queries for a fix.
Ran out of time and overwhelmed, I restored backups and buried my head in the sand. Nope, not now. Future, yes, but oh not now.
Sometimes we get so engrossed in what we're doing we can't see the problem(s). I do that a lot, so I have take a break. Same with creating music. You get so deaf to what you are trying to write that nothing sounds good no matter what you do. In the words of Snoop Dog, 'I had to back up off of it and sit my cup down. Tanqueray and chronic, yeah, I'm fucked up now.'
Take a break.
I am currently arguing what to do with my gaming rig and home theater. Either get a long cable which would need a DP-to-HDMI adapter or get a used mini PC (which is currently cheaper than a Raspberry Pi?) and setup Sunshine and Moonlight (but over WiFi and not LAN) to be more flexible when I eventually move the two into separate rooms. Does anyone have some experience with that? Maybe also latency over wireless network?
I've just set up Wireguard, so I can access my home network from everywhere, but the old laptop that I wanted to use as a server has just quit. So now I have to find a different machine
I've been fending off AI bots the last week or so; wrote about it here:
https://gerowen.substack.com/p/the-ai-data-scraping-is-getting-out
I have a self-hosted AI system that works pretty well. I can interact with it via my phone, the shell, my IRC server, and I can verbally talk to it.
But I want to get it to remember things, so I need to start working on RAG or something. Eventually I'd like to be able to have it draft emails for me, and schedule appointments.
Shoutout to @Estebiu@lemmy.dbzer0.com for helping me appreciate the joy of docker compose. I got to set up Navidrome and it's been great!
With that said, I have a security-related question: at what point in self-hosting am I exposed to the outside internet that warrants things like reverse proxies and other security measures? I'm currently typing router IPs (e.g. 192.168.x.x) to access the services, so is my machine exposed if the only people intending to connect are local on our wireless network?
To expose your stuff to the outside internet, you need to actively set port forward in your internet router, you won't do that by accident.
Building a simple workflow with AI agent for our community watch group. Also building an open source automation platform, currently working through GUI templates for it.
A new homepage for the business of my wife.
I plan to use Hugo for it, I just wish the documentation would be better.
For the homepage I need a few additional "non-blog" pages and from the documentation I am not sure how to do that the best way.
But to be honest, I have not really looked deeper into that, so it is very possible that I just missed something.
My radarr instances won't download anything. It will search and find compatible torrents, but then it just spins and spins, nothing ever moves to the queue. If I refresh its like nothing happened at all. I confirmed that qbt is running properly and my Sonarr instances seem to be running ok.
I recently reorganized the root files to separate HD/UHD content so that I can run 2 instances for Overseerr requests, then this issue started. I had to reset the root folders and now there's also a root folder error about collections that I can't resolve either... got me thinking about doing a full reinstall.
i run coolify and I have to make my own solutions so I'm learning a lot about docker.