sorted by: new top controversial old
[-] Guenther_Amanita@slrpnk.net 2 points 2 days ago

I also made a very similar comment, but with uBlue (Bazzite, Aurora, Bluefin) instead.

They are still pretty vanilla, but include a big list of QoL stuff added in, like staged updates, Distrobox, a huge list gaming tweaks in Bazzite, and much more.

It's basically stock Atomic made right!

I've used them for a year now, and they're fantastic!

Just a small heads up for OP: You have to do quite a lot of (advanced) things differently from now on if you choose Atomic. Use containers (Distrobox, etc.) for everything you can, avoid installing stuff on the host if possible, etc.

[-] Guenther_Amanita@slrpnk.net 12 points 2 days ago* (last edited 2 days ago)

Because others already suggested Arch/ EndeavourOS, I'll be suggesting something else: Bazzite.

It's part of the image based ("immutable") Fedora series and is basically Fedora Kinoite, with all drivers and codecs already set up for you, self managing, with many gaming tweaks included.

It's rock solid and basically unbreakable, while also being extremely modern and updated. On Arch, even if it doesn't break, you always get the newest stuff, which might not be as polished. On Fedora, it matures a few months, while still being very modern.

The main target group is "For Linux users who don't want to use Linux", meaning, it runs all your favourite stuff (KDE, etc.) without having to care for anything. It even updates itself automatically in the background without any interference.

If you prefer something with less "bloat" (a lot of optional tools and software to choose from, but nothing mandatory), then check out Aurora, which is basically the same, but without gaming stuff.

For more information, check out universal-blue.org

Just a small heads up for OP: You have to do quite a lot of (advanced) things differently from now on if you choose Atomic.

Use containers (Distrobox, etc.) for everything you can, avoid installing stuff on the host if possible, etc.

Just use Flatpaks for 95% you do graphically, and for CLI stuff or software that isn't available as Flatpak, I would recommend you to create an Arch Distrobox container (already set up IIRC) and use that. You can even install stuff from the AUR and export it, so it works just like it is supposed to.

[-] Guenther_Amanita@slrpnk.net 5 points 3 days ago* (last edited 3 days ago)

Maybe, another consideration might be to not run Linux on Windows in some way, but the other way around.

Linux offers great virtualization, maybe you can use QEMU with KVM and GPU passthrough, and then run Windows inside this box.

I find Linux more powerful and less annoying to use day to day, and having those annoyances inside a small virtualized container I can just shut down is more peaceful.


WSL can be restricting, since Linux can't access anything, and I think getting "the real and proper thing" might be better.

And dual booting, by having both Windows and Linux on the same drive, is something I would advise against. Windows doesn't play nice with others and often "accidentally" breaks the bootloader and hard drive permissions, leading only to trouble. If you dual boot, install them on a separate drive and select the booting drive manually in the BIOS.


Also, why do you want to run Ubuntu specifically? Did you also look up for alternatives, like Fedora or Debian?

[-] Guenther_Amanita@slrpnk.net 9 points 3 days ago

Dual booting Pros: a full-fledged Linux OS Cons: Harder to install and to mantain.

Also, sometimes Windows being an ass and "accidentally" breaking the bootloader.

I advice anyone to have just one OS per drive installed. Keep Windows and Linux separate if possible, or some Windows update may break GRUB.

[-] Guenther_Amanita@slrpnk.net 8 points 1 week ago

I use Casa"OS". It's fine, but nothing groundbreaking. Cockpit for example can do pretty much the same, and for Docker containers, I nowadays mostly use docker compose.

But hey, it helped me a quite a bit in the beginning, and it's cool. Pretty basic, but enough for most people, mainly beginners.

[-] Guenther_Amanita@slrpnk.net 44 points 1 week ago* (last edited 1 week ago)

Then how do you explain the continued success of Mint?

Because Mint's philosophy is to make a friendly, simple and usable system for everyone.

That may be for people who came from Windows before, or those who like their OS to be a bit more conservative, meaning no flashy stuff, boring, and just working. Just like Windows was "in the good ol' days".

This makes it accessible and usable by everyone, including Linux sysadmins who come home after work and don't want to deal with annoying computers and fixing things.

Everything on Mint feels high quality, functional and cohesive.

ElementaryOS on the other hand feels like a cheap MacOS clone, but nothing works. Those who want Mac, buy a Mac.

Mint/ Cinnamon on the other hand is similar to Windows (XP, 7, etc.), but not a copycat. It's familiar enough to be intuitive for Windows users, but much enough it's own thing.

Mint's main focus is to get a uncomplicated, and usable system, while Elementary's focus is to just do what Apple does. ... Well, did. 15 years ago. They totally forgot how much work maintaining a distro and a desktop with a whole app suite is, and just stopped working on it.

While Gnome and KDE (and other WMs/ DEs) got magnitudes better in just one year (e.g. Plasma 6), Pantheon (and Elementary) just stagnated the last 5 years or so.

They don't even offer/ work on Wayland yet, or other new things.

Either they'll stop working on Elementary, and focus only on Pantheon, so it can live on on other distros, or it will just continue dying like it does currently.

[-] Guenther_Amanita@slrpnk.net 4 points 1 week ago

Logseq and Obsidian are only similar on the first look, but very different usage wise. Both are very open with a plugin system, and you can modify them to turn them into one eachother.

So, if you want only FOSS, then Logseq is the only choices you have.

But Obsidian is, even though it's proprietary, very sane. Open plug-in system, active community, great devs who don't have much against FOSS, and more.


Obsidian

  • More similar to a classic note taking app, like OneNote, but with a lot of features. Hierarchical structure, and more of an "essay" style, where you store a lot of text in one page.
  • Page linking is only done when you think it makes sense
  • Has been a bit longer around than Logseq, feels more polished
  • Great sync and mobile app, which support plugins from what I've heard

Logseq

  • Non-linear outliner. Every page is on the same level, but within a text passage, the indentation matters (parent-child-relationship)
  • You create a LOT of more pages. Most of my pages are empty. They are mainly there for linking topics. I rarely create pages manually.
  • The journal is where you write most stuff. You then link each block to a page.
  • Logseq a bit "special". May not be for everyone. I for example am a bit of a disorganised thinker, who mentally links a lot of knowledge and throws concepts around all the time. Logseq is my second nature, because it's more flexible. My GF on the other hand is more structured, and prefers something like Apple Notes, or, if she would care about note taking, something like Obsidian.
  • The mobile app isn't great. It's fine when I'm not at home, but the desktop version is the "proper" one, and mobile/ iPad a second class citizen.
  • Sync is only experimental for now. It will soon be officially supported (hopefully) and self hostable, but it worked fine for me.
[-] Guenther_Amanita@slrpnk.net 3 points 1 week ago

I don't see any problems with that. Even I (and probably most others here), who are FOSS advocates, think Obsidian's model is fine.

The devs surely get why FOSS is important, and try their best to match the pros of open source. They even stated that if the company goes bankrupt or they stop developing the app, they'll open source it.

One major thing they do absolutely right is how the notes get stored. On other note taking apps, it's a proprietary database, often "in the cloud", where your notes get hold hostage. Here, they're just Markdown files, and the whole thing is pretty open, encouraging a strong community.

It's similar to Valve/ Steam. Proprietary, but liked by most Linux people.

[-] Guenther_Amanita@slrpnk.net -1 points 1 week ago

That's not what OP was asking for though.

Why not? In my comment I explained exactly what benefits it would have in this case with a Nvidia GPU. I think it makes sense to at least mention the option.

OP tried Bazzite and wasn't the biggest fan of it, but not because it's image based, no, just because it uses the same Nvidia driver as upstream Fedora.

They could also have said that they really liked it, who knows?

And I don't understand why the Linux user community on Lemmy pushes immutable distros so hard.

Because they're awesome? They're extremely low maintenance, just work (for me), are very robust, offer a lot of choice, and much much more.

I think they're very underrated and should be used much more. Sure, some people just don't like them, but some people would, and those should know this option exists.

[-] Guenther_Amanita@slrpnk.net 2 points 1 week ago

I personally think X11 shouldn't be used anymore. Fedora dropped official support for it recently iirc and it will soon be deprecated, so it might be even worse in the future.

Wayland works perfectly fine under Gnome from what I've heard, and with Plasma, it should be working great too.

[-] Guenther_Amanita@slrpnk.net 3 points 1 week ago* (last edited 1 week ago)

You could maybe try Bazzite or Aurora/Bluefin.

They are all Fedora Atomic, the "immutable" Fedora variant, and offer baked in Nvidia support.

The cool thing is:

  1. If the driver/ Wayland breaks on your install, then it will break on thousands of others simultaneously, and the devs can fix it very very quickly, because every installation is identical.
  2. If it breaks, you can roll back in seconds and keep using the image that still worked yesterday. And in the meantime, the developers are already working on a fix, which takes just hours or a day max.
  3. You don't have to install and update anything yourself. Just do your computer stuff and stop worrying.
  4. There's also a GTS (or whatever it's called) variant around, which is the last major version of Fedora. You won't get the newest stuff and will be half a year behind in terms of features, but then there won't be any surprises. I believe the bluefin:gts isn't around yet, but will come with the next major release.
[-] Guenther_Amanita@slrpnk.net 11 points 1 week ago

You can maybe post something on local marketplaces. Something like Kleinanzeigen (Germany), Facebook Marketplace (US I think) or something similar. Or local groups.

There are a lot of hobbyists who have a printer that stands still for 99% of the time, who can print it for cheap.

Or, as someone else already has suggested, your local library.

246
submitted 3 weeks ago* (last edited 3 weeks ago) by Guenther_Amanita@slrpnk.net to c/3dprinting@lemmy.world

I don't print any abrasive materials at all. Pretty much only normal PLA and PETG.

I noticed, that my print quality gradually went down quite a bit, especially in the last few prints. I had a lot of stringing, weird blobs, and scarred surfaces.

Now, the print quality is as good as it should be!

They are dirt cheap. You can get a set of 10-15 generic ones, in different sizes, for only a few bucks. Don't forget that they are consumables.

30
submitted 1 month ago* (last edited 1 month ago) by Guenther_Amanita@slrpnk.net to c/selfhosted@lemmy.world

TL;DR:

  • I can't decide between Debian and the new "immutable" Fedora server variants
  • Currently I use Debian with pretty much everything being containerised, and it works fine.
  • I'm neither very good at what I'm doing, nor want to spend my weekends troubleshooting. Opting for something new could cause some headaches I guess?
  • How did you set up CoreOS? Are there simple ways?
  • Would you recommend me something different?

My backstory with Debian

I will soon set up a new home server and need your opinion and experiences.

I'm using Debian as the OS for my current one.
While it doesn't match my "taste" perfectly, as I slightly prefer RedHat stuff, I really don't have much preference, since I don't interact with the host much anyway.
Everything is containerised via Docker, and I don't even know why I like Rocky-/ Alma more. I tried Alma once and it just clicked better, I can't explain it...
But that doesn't mean I dislike Debian, not at all!

Still, at that time I decided to go with Debian, since it's the standard for most selfhosters, has the best software support, and is completely community run, opposed to RHEL and its clones.

At that time I didn't know Distrobox/ Toolbx, and I really wanted to install CasaOS (basically a simplified Cockpit + Portainer for less techy people), because I was a total noob back than and didn't want to do everything via CLI.

Nowadays, I found alternatives, like Cockpit, and I also do more via the terminal.
And if I want to install something that doesn't support my host OS, then I just enter my Toolbx and install it there.

Still, I absolutely don't regret going for Debian. It was a good choice. It's solid and doesn't get in my way.


What has changed in the last year(s)

In the last year now, I really began to enjoy using image based distros, especially Fedora Atomic.
I really love Atomic as desktop distro, because it is pretty close to upstream, while still being stable (as in how often things change).

For a desktop workstation, that's great, because DEs for example get only better with each update imo, and I want to be as close to upstream as possible, without sacrificing reliability, like on a rolling release.
The two major releases each year cycle is great for that.

But for a server, even with the more stable kernel that's used in CoreOS from what I've heard, I think that's maybe too unstable?

I think Debian is less maintenance, because it doesn't change as often, and also doesn't require rebooting after each transaction.

But, on the contrary, I wouldn't loose much to the "immutability", because I use containers for everything anyway.
Having way better security (sane SELinux setup, rootless containers, untampered OSA, etc.) and the ability to roll back in case something doesn't work, while self updating, sounds very promising.


Setting up CoreOS; FCOS vs FIOT

The major thing that's keeping me away from CoreOS/ uCore is all the ignition-butane-stuff.
From what I've heard, it's needlessly complicated for home use, and FCOS is best suited for fleets/ clusters of servers, not just for one.

Fedora IOT seems to be simpler, but doesn't have the same great defaults and features as uCore, since there isn't an IOT variant of uBlue.
But hey, at least I have my Anaconda installer.

What do you think about installing IOT, and then rebasing to uCore?
Or, do you think FCOS is just not the right thing for my use case?

In general, do you think that it is worth it, compared to plain old Debian?


Pros vs. cons

Anyway. I'm really thinking about all of this for a long time now, and can't decide.

On the one side, it all sounds promising and great.
But, on the other side, selfhosting isn't a primary hobby of mine. I just want a solid setup I don't have to maintain much after setting everything up. Image based server OSs are still very new and often unheard of, and being an early adopter might cause a lot of headache in that case when it comes to servers.


The "right" use case?

Just in case no one has tried FCOS or FIOT here, I will continue using Debian for my main server, and only use Fedora IOT for my Octoprint server, which only gets turned on sporadically, and would greatly benefit from that.

But if there are positive experiences, then I might give it a try.


Alternatives

Or, would you recommend me something entirely different?

NixOS for example sounds great in theory, but is way too complicated for me personally.

Or, would you recommend me to give Alma another try?

Is there something even better?

8

First of all, thank you so much for your great answers under my post from yesterday! They were really really helpful!

I've now decided that I will not use something with USB. It really doesn't seem to be reliable enough for constant read-write-tasks, and I don't wanna risk any avoidable data loss and headache.

Also, it just doesn't seem to be very future proof. It would be pretty expensive, only for it to get replaced soon, and then getting obsolete. It just seemed like a band-aid solution tbh. So, no USB hard drive bay, no huge external hard drive, and no NAS just for that purpose.


A few people asked me about the hardware.

My server is a mini-PC/ thin client I bought used for 50 bucks. I've used it for about two years now, and it had even more years of usage under the belt with its' former owner. Imo, that's a very sustainable solution, that worked pretty well until now.

I used it almost exclusively for Nextcloud (AIO), with all the data being stored in the internal 1 TB SSD.

For those who are interested, here are all the hardware details:

<hwinfo -short>

Thing is, I want to get more into selfhosting. For that, my main goal is to
a) Replace Nextcloud with individual (better) services, like Immich and Paperless-ngx.
NC-AIO was extremely simple to set up and worked pretty fine, but I always found it to be bloated and a bit wonky, and, mainly, the AIO takes up all my network and resources. I just want something better, you understand that for sure :)
b) Get more storage. I'm into photography, and all those RAW photos take up SO MUCH SPACE! The internal 1 TB is just not future proof for me.
c) Maybe rework my setup, both in software, and maybe in hardware. Originally, I didn't plan to screw everything, but I think it might be better that way. The setup isn't bad at all, but now, as I got more experience, I just want it to be more solid. But I'm not sure about doing that tbh, since it really isn't a lost case.


As someone already mentioned in the last post, I really don't have a million bucks to create my own data center. I'm not completely broke, but almost :D
Therefore, I just want to make the best out of my already existing hardware if possible.

Because I decided against USB, and because I don't know if there are any slots on the mainboard that can be repurposed for additonal storage, I need your advice if there are any options to achieve that, e.g. via a PCIe slot + adapter, if I had any.
I saw one SATA III port, but that one really isn't enough, especially for extendability.

Here are the photos from both the front and back side:


My thought was, instead of buying one hella expensive 3+TB SSD drive, just screw it and make something better from scratch.

So, if you guys don't give me a silver bullet solution, aka. "you can use this slot and plug in 4 more drives", I will probably have to build my own "perfect" device, with a great case, silent fans, many storage slots, and more.

Btw, do you have any recommendations for that? (What mainboard, which case, etc.) Preferably stuff that I can buy already used.

Thank you so much!

56
submitted 1 month ago* (last edited 1 month ago) by Guenther_Amanita@slrpnk.net to c/selfhosted@lemmy.world

I'm planning to upgrade my home server and need some advice on storage options. I already researched quite a bit and heard so many conflicting opinions and tips.

Sadly, even after asking all those questions to GPT and browsing countless forums, I'm really not sure what I should go with, and need some personal recommendations, experience and tips.

What I want:

  • More storage: Right now, I only have 1 TB, which is just the internal SSD of my thin client. This amount of storage will not be sufficient for personal data anymore in the near future, and it already isn't for my movies.
  • Splitting the data: I want to use the internal drive just for stuff that actively runs, like the host OS, configs and Docker container data. Those are in one single directory and will be backed up manually from time to time. It wouldn't matter that much if they get lost, since I didn't customize a lot and mostly used defaults for everything. The personal data (documents, photos, logs), backups and movies should each get their own partition (or subvolume).
  • Encryption at rest: The personal data are right now unencrypted, and I feel very unwell with that. They definitely have to get encrypted at rest, so that somebody with physical access can't just plug it in and see all my sensitive data in plain text. Backups are already encrypted as is. And for the rest, like movies, astrophotography projects (huge files!), and the host, I absolutely don't care.
  • Extendability: If I notice one day that my storage gets insufficient, I want to just plug in another drive and extend my current space.
  • Redundancy: At least for the most important data, a hard drive failure shouldn't be a mess. I back them up regularly on an external drive (with Borg) and sometimes manually by just copying the files plainly. Right now, the problem is, if the single drive fails, which it might do, it would be very annoying. I wouldn't loose many data, since they all get synced to my devices and I then can just copy them, and I have two offline backups available just in case, but it would still cause quite some headache.

So, here are my questions:

Best option for adding storage

My Mini-PC sadly has no additional ports for more SATA drives. The only option I see is using the 4 USB 3.0 ports on the backside. And there are a few possibilities how I can do that.

  • Option 1: just using "classic" external drives. With that, I could add up to 4 drives. One major drawback of that is the price. Disks with more than 1 TB are very expensive, so I would hit my limit with 4 TB if I don't want to spend a fortune. Also, I'm not sure about the energy supply and stability of the connection. If one drive fails, a big portion of my data is lost too. I can also transform them into a RAID setup, which would half my already limited storage space even more, and then the space wouldn't be enough or extendable anymore. And of course, it would just look very janky too...
  • Option 2: The same as above, but with USB hubs. That way, I theoretically could add up to 20 drives, when I have a hub with 5 slots. That would of course be a very suboptimal thing, because I highly doubt that the single USB port can handle the power demand and information speed/ integrity with that huge amount of drives. In reality, I of course wouldn't add that many. Maybe only two per hub, and then set them up as RAID. That would make 4x2 drives.
  • And, option 3: Buy a specialized hard drive bay, like this simpler one with two slots or this more expensive one for 4 drives and active cooling. With those, I can just plug in up to 4 drives per bay, and then connect those via USB. The drives get their power not from the USB port, but from their own power supply. Also, they get cooled (either passively via the case if I choose one that fits only two drives, or actively with a cooling fan) and there are options to enable different storage modes, for example a built in RAID. That would make the setup quite a bit simpler, but I'm not sure if I would loose control of formatting the drives how I want them to be if they get managed by the bay.

What would you recommend?

File system

File system type

I will probably choose BTRFS if that is possible. I thought about ZFS too, but since it isn't included by default, and BTRFS does everything I want, I will probably go with BTRFS. It would give me the option for subvolumes, some of which are encrypted, compression, deduplication, RAID or merged drives, and seems to be future proof without any disadvantages. My host OS (Debian) is installed with Ext4, because it came like that by default, and is fine for me. But for storage, something else than Ext4 seems to be the superior choice.

Encryption

Encrypting drives with LUKS is relatively straight forward. Are there simple ways to do that, other than via CLI? Do Cockpit, CasaOS or other web interface tools support that? Something similar to Gnomes' Disk Utility for example, where setting that up is just a few clicks.

How can I unlock the drives automatically when certain conditions are met, e.g. when the server is connected to the home network, or by adding a TPM chip onto the mainboard? Unlocking the volume every time the server reboots would be very annoying.

That of course would compromize the security aspect quite a bit, but it doesn't have to be super secure. Just secure enough, that if a malicious actor (e.g. angry Ex-GF, police raid, someone breaking in, etc.) can't see all my photos by just plugging the drive in. For my threat model, everything that takes more than 15 minutes of guessing unlock options is more than enough. I could even choose "Password123" as password, and that would be fine.

I just want the files to be accessible after unlocking, so the "Encrypt after upload"-option that Nextcloud has or Cryptomator for example isn't an option.

RAID?

From what I've read, RAID is a quite controversial topic. Some people say it's not necessary, and some say that one should never live without. I know that it is NOT a backup solution and does not replace proper 3-2-1-backups.

Thing is, I can't assess how often drives fail, and I would loose half of my available storage, which is limited, especially by $$$. For now, I would only add 1 or max 2 TB, and then upgrade later when I really need it. And for that, having to pay 150€ or 400€ is a huge difference.

57
submitted 2 months ago* (last edited 2 months ago) by Guenther_Amanita@slrpnk.net to c/linux@lemmy.ml

Don't get me wrong. I absolutely love Fedora Atomic (Silverblue, Bazzite, Kinoite, Aurora, IOT, etc.), more than any other distro I used, and I plant to continue using it.

It never made any problems on any of my devices, and because it is pretty much indestructible and self-managing, I even planned to install it on my Mum's new laptop, in case her current one (basically a toaster with Mint on it) breaks.

But with the last days, my trust is damaged quite a bit.

First one, where I couldn't update anymore on uBlue, because of faulty key pairs. This is a huge thing for me because uBlue updates in the background, and if I wouldn't have read it here on Lemmy, I would have found out way too late, which is a security risk imo.

And now, my devices weren't able to boot anymore due to some secure boot stuff. Again, if I wouldn't have subscribed the Fedoramagazine, I would have noticed it way too late.
I was able to just boot into an older image and just paste a few commands from the magazine's post, and it was resolved in just seconds (download time not included).

Both instances were only a minor thing for ME.
But both would have been a headache if I wouldn't follow those blogs, which is a thing only nerds (like myself) do.
Nobody else cares about their OS, it is supposed to just work, hence why I use Atomic.

I don't wanna blame the devs (both j0rge/ uBlue and the Fedora team), they were very quick, transparent and offered very simple fixes.
And, being able to just boot into an older image, just in case, is something I am very thankful for, but nothing I want to depend on.

Having to be informed about stuff like this and then having to use the CLI is just a no-go for most people.

Am I over-reacting about this too much? What's your view on those things?

view more: next ›

Guenther_Amanita

joined 4 months ago