I just map both the user cache and the /tmp directory to a RAM drive. I allocated 4 GB but in practice it never gets even close to that much, and Linux seems to not be reserving the entire 4 GB at boot so I would assume how much RAM is used depends on how much is actually in your cache.
It also defers cache and tempfile related problems to turning it off and on again.
seems like a bug in one of rhe programs you’re using.
modt software automatically manages it’s cache…
are you using build caching tools such as Mozilla sccache? These tend to create 20gb+ cache directories, especially if used with debug builds
My only issue is software availability and management. I use the Packman repository to manage codecs and I avoid using the change vendor option; i used to change the vendor every time and ended up with a broken system, so I reinstalled and also resized my partition because I dual boot. I haven’t had problems at all.
You only need to pay attention for your needs, I recently installed systemd networking packages because they don’t come preinstalled, and YaST is very helpful in some situations like installing patterns (multiple related packages at once), mostly desktop environments. I gotta say that the openSUSE Wiki may not be enough to understand, but there is an official forum and you can also look at the Arch wiki.
Btw, GNOME is the official DE used by the developers, but KDE Plasma works very well, and all of them update constantly, you’ll have available updates every week.
I haven’t used it yet personally, but I would bet as soon as Debian/Ubuntu LTS/CentOS/openSuse/other stable Linux distros get kernels new enough to it will be not just a btrfs killer, but a ZFS killer too.
The tiered write and read layers and SMR support put ZFS caching to shame.
Nobody uses SMR for live data anyway unless it’s in very particular circumstances.
Bcachefs is still at least a couple of years away from serious use. But sure, if it’s available and you have a good backup strategy you can use it today.
As for “years away” I agree. As my first post said people should wait till you can use bcachefs in the stable distros. Debian isn’t getting kernel 6.7 any time soon 😆. So years away is right in any case.
I think bcachefs addresses the reason why people don’t use SMR HDDs. (Aka changes resulting in cascading writes)
You could have a data pool with the following tiers.
Tier 1: SSDs
Tier 2: HDDs
Tier 3: SMR HDDs
With bcachefs you would only ever write to your tier 1 storage. In the background, as able, bcachefs would offload the data from the faster lower tiers to the slower higher tiers based on frequency of data access.
You would only ever read from the SMR HDDs and would never write to them. They act as a sort of async backing to your data.
Personally I would love a data pool with a few SSDs, backed by a few HDDs, backed by many SMR HDDs. You would save so much money just with good architecting.
Bcachefs should be a ZFS killer. All the features of ZFS with storage tiers being a superior version of ZFS’s L2arc with none of the DKIM kernel license incompatibility nonsense.
Damn, I didn’t think to include SMR drives when it comes to bcachefs. Your whole comment made me appreciate the whole concept under a whole new light actually, thanks!
Let me put it like this: it’s about learning curve. Arch is relatively easy to begin with, but NixOS gets much easier the more nix you learn.
What do I mean about that? Imagine having to patch something, which can be the thing. On arch you’d have to replace a package, which could lead to issues and conflicts, whereas NixOS gives you the option to keep two or even more versions of the same library, because it does not rely on your traditional UNIX path.
But with this super power comes a catch. You have to learn a programming language and learn how the nix store operates, which is a pretty high learning curve. Also, NixOS suffers from a governance issue and going by the documentation is like shooting in the dark.
That being said, the best manual for NixOS is GitHub, searching for anything and filtering by the nix language. You’ll see a ton of varying systems, be they workstations or servers.
And no matter what all the warnings say, no, flakes aren’t EXPERIMENTAL or UNSTABLE, but rather CONTENTIOUS internally. Again: I love NixOS, but they gotta fix their governance issues.
I don’t use those two flags, but have several pis running docker with no issues. They’ve been running (almost) 24/7/365 going on maybe 2 years now with the same sd cards.
Probably a good idea to look for a different client, call me tinfoil but I wouldn’t want to touch a very old mechanism that is supported/pushed by a very recognisable 3 letter agency
A surprising amount of services (including Azure last I tried) can only handle RSA keys, so after trying ecdsa only for a while I ended up adding a RSA key again.
With that said - it’s 2023, in almost all cases you should have your keys in a hardware module nowadays, in which case you’d use a different command for keygeneration.
Actually it is the same story with TLS 1.3 and TLS 1.2. A bunch of sites still doesn’t support TLS 1.3 (e. g. arstechnica.com, startpage.com) and some of them only support TLS 1.2 with RSA (e. g. startpage.com).
You can try this yourself in Firefox by disabling ciphers (search for security.ssl3 in about:config) or by setting the minimum TLS version to 1.3 (security.tls.version.min = 4 in about:config).
Strange enough TLS 1.3 still doesn’t support signed ed25519 certificates :| P‐256, NIST P‐384 or NIST P‐521 curves are known to be “backdoored” or having deliberately chosen mathematical weakness. I’m not an expert and just a noob security/selfhoster enthusiast but I don’t want to depend on curves made by NSA or other spy agencies !
I also wondering if the EU isn’t going to implement something similar with all their new spying laws currently discussed…
linux
Oldest
This magazine is from a federated server and may be incomplete. Browse more on the original instance.