Keep notes on what you do including outcomes. You can always reinstall, skip all of the extra crap you didn’t need to mess around with and have a good clean system without having to back track.
It’s easy to forget the steps you took to do something on your computer, especially several months later when you’re trying to upgrade. Sometimes when you try several different ways of solving a problem, it’s easy to forget which method was successful the next day!
It’s a very good idea to take notes about the software you’ve installed and configuration changes you’ve made. When editing configuration files, it’s also a very good idea to include comments in the file explaining the reason for the changes and the date they were made.
This has saved me so much grief. If ever I mess up a system so badly that I want to re-install or when I want to set up a new machine, having a clear set of notes makes it a breeze.
It’s easy to forget the steps you took to do something on your computer, especially several months later when you’re trying to upgrade. Sometimes when you try several different ways of solving a problem, it’s easy to forget which method was successful the next day!
Hello OP, warm welcome to the schizophrenic community that is linux! I’m running this exact same setup as you intend to.
Couple of points I’d add:
Nvidia and linux is a shit show. You either use the gimped mesa drivers (not so good), or the closed source drivers (even worse), especially on Wayland. If you still want to try it out, I suggest you stick with X11. I was using a 3080Ti and gave up trying to get it to work on Wayland+KDE Plasma v5 without screen glitches and bought a 7900XTX instead. AMD works out of the box without further configuration. I’ve had nothing but issues with Nvidia. For the GPU neckbeards that are going to achytually me, please don’t, I have no skin in this game as I have a bunch of cards from both.
I am not a pro nor expert in the foundations of linux, but I more or less know my way around its fundamentals (enough not to get into too much trouble), so I would actually suggest not getting into a rolling-releasing distro like Tunbleweed until you are super comfortable with a little hands-on and figuring shit out. Pop!_OS might be a better bet for your use case for now because it comes bundled with Nvidia drivers, but it uses a GNOME derivative (cosmic DE) so you don’t actually get the KDE experience.
Having said these, I absolutely adore Tumbleweed and KDE, I’ve been half daily-driving it but the gaming experience is not the absolute best (be prepared to experience weird glitches and crashes). If a flawless gaming experience is non-negotiable to you, stick with Windows. But if you’re ready to explore the quirky wonders of linux, the beauty of it all is the experience and the real reward is the friends you make along the way.
The hardest part is when you need to use tools from windows, or to develop a program for that system. Also, many things require good amount of configuration, and with that, expertise at what you doing. I’m currently struggling with solving bug as to why VSCodium will not debug in external terminal.
A biggie you miss is the toolchain: the compiler/binutils/linux-headers/libc/libstdc++ combination. The libc and usually libstdc++ are key components of any install. The other parts usually don't make it to non-dev-desktops, but the distro couldn't be made without them, so they're virtually always available as packages.
Only exception is if the entire distro is cross-compiled or it's made exclusively for containers, but those kinds of special distros break every rule imaginable anyway. Some might not even ship a bootloader or a Linux kernel by themselves.
I really like these suggestions, I’ve always wanted to contribute to FOSS software, but always felt underskilled. I will add this to my list of things to do to challenge my Linux and basic programming skills.
It’s worth noting that the barrier to entry as a maintainer depends on which distro you’re using at the time. It’s not uncommon for a distro to have a community repository system, like PPAs in Ubuntu, AUR for Arch, MPR for Debian, etc. I’m not very familiar with Mint, and couldn’t easily tell if it has its own or just uses PPAs from upstream.
It isn’t especially taxing on programming skills, and if you don’t pick too complex of a package, the Linux skills required shouldn’t be wildly above your level, but may push you to learn some new things by digging a bit deeper. I haven’t formally maintained public packages, but I’ve needed to build a few over my years using Linux, and it was easier than I’d expected to just build one. It may be easier than you think, too.
Thanks for the additional info. I installed Arch, it was far less daunting then I anticipated. In fact, it was prettt straightforward. I’ll look into your suggestion.
It seems like what you’re asking about are more what I’d think of as components of an a Linux “system” or “install.”
First off, it’s definitely worth saying that there aren’t a lot of rules that would apply to “all” Linux systems. Linux is huge in embedded systems, for instance, and it’s not terribly uncommon to find embedded Linux systems with no shells, no DE/WM, and no package manager. (I’m not 100% sure a filesystem is technically necessary. If it is, you can probably get away with something that’s… kinda sorta a filesystem. But I’ll get to that.)
Also, it’s very common to find “headless” systems without any graphical system whatsoever. Just text-mode. These are usually either servers that are intended to be interacted with over a network or embedded systems without screens. But there are a lot of them in the wild.
There’s also Linux From Scratch. You can decide for yourself whether it qualifies as a “distribution”, but it’s a way of running Linux on (typically) a PC (including things like DE’s) without a package manager.
All that I’d say are truly necessary for all Linux systems are 1) a bootloader, 2) a Linux kernel, 3) A PID 1 process which may or may not be an init system. (The “PID 1 process” is just the first process that is run by the Linux kernel after the kernel starts.)
The “bunch of default applications and daemons” feels like three or four different items to me:
Systemd is an example of an “init system.” There are several available. OpenRC, Runit, etc. It’s main job is to manage/supervise the daemons. Ensure they’re running when they’re supposed to be. (I’ll mention quickly here that Systemd has a lot more functionality built in than just for managing daemons and gets a bad rap for it. Network configuration, cron, dbus for communication between processes, etc. But it still probably qualifies as “an init system.” Just not just an init system.)
Daemons are programs that kindof run in the background and handle various things.
Coreutils are probably something I’d list separately from user applications. Coreutils are mostly for interacting with low-ish level things. Formatting filesystems. Basic shell commands. Things like that.
User applications are the programs that you run on demand and interact with. Terminal emulators, browsers compilers, things like that. (I’ll admit the line between coreutils and user applications might be a little fuzzy.)
As for your question about graphical systems, X11 and Wayland work a little differently. X11 is a graphical system that technically can be run without a desktop environment or window manager, but it’s pretty limited without one. The DE/WM runs as one or more separate processes communicating with X11 to add functionality like a taskbar, window decorations, the ability to have two or more separate windows and move them around and switch between them, etc. A Wayland “compositor” is generally the same process handling everything X11 would handle plus everything the DE/WM would handle. (Except for the Weston compositor that uses different “shells” for DE/WM kind of functionality.)
As far as things that might be missing from your list, I’ll mention the initrd/initramfs. Typically, the way things are done, when the Linux kernel is first loaded by the bootloader, it an “initial ramdisk” is also loaded. Basically, it creates a filesystem that lives only in ram and populates it from an archive file called an “initramfs”. (“initrd” is the older way to do the same thing.) Sometimes the initramfs is bundled into the same file as the kernel itself. But, that initial ramdisk provides an initial set of tools necessary to load the “main” root filesystem. The initramfs can also do some cool things like handling full disk encryption.
So, the whole list of typical components for a PC-installed Linux system to be interacted with directly as I’d personally construct it would be something like:
Bootloader
Linux Kernel
Initramfs
Filesystem(s)
Shell(s)
Init system
Daemons
Coreutils
Graphical system (X11 or Wayland potentially with a DE/WM.)
User applications
Package manager
But techinically, you could have a functional, working “Linux system” with just:
Bootloader
Linux Kernel
Either a nonvolatile filesystem or initrd/initramfs (and I’m not 100% sure this one is even strictly necessary)
A PID 1 process
Hopefully this all helps and answers your questions! Never stop learning. :D
You would need some non volatile storage to hold your bootloader be that on the network or local. Also any shell more complicated than tty will need to store config files to run.
Systemd has gone way and beyond what was supposed to be a replacement for init.rc.
Most important thing… not ALL Linux distros include systemd as the default init system. That’s the beauty of Linux (and POSIX in general), you can choose.
If I understand correctly, the filesystem driver is contained within the kernel for all linux-native filesystems (Ext4, XFS, BtrFS, F2FS, etc.), just as drivers for computer components and devices are. But drivers to access NTFS (Windows) and HFS+ (Mac OS) drives are programs in userspace
linux
Hot
This magazine is from a federated server and may be incomplete. Browse more on the original instance.