<span style="color:#323232;">Ctrl-z
</span><span style="color:#323232;">kill -9 %1 # Shell keeps track of job pids for you, job 1 is %1, job 2 is %2, etc
</span><span style="color:#323232;">fg # Not technically necessary, but it's fun to see the corpse
</span>
I think the idea of a (relatively) simple or as complex “roll your own flavour” OS makes lots of sense to someone like me. For most people the effort might not be worth the payout.
you can create them afterwards and move the stuff into the subvol. do it from a live usb and don’t forget to update fstab. be sure to use rsync with the flag to keep permissions etc
That’s simple, but it’s a completely unnecessary waste of I/O. You could create a writable snapshot of the btrfs root as a subvolume, edit the fstab and any other relevant files within that new subvolume, reconfigure the bootloader to specify that subvolume as the root filesystem (as a Linux kernel command line argument) instead of the btrfs root, and then reboot. After rebooting, the original btrfs root can be mounted, and everything unwanted from the original root (other than the new subvolume and its ancestor directories, obviously) can be deleted. Do not delete anything that you didn’t want to lose the changes to on the original root subvolume that you did after creating the snapshot, as the snapshot only remembers what you did before, as well as the changes made specific only to it (like the fstab).
If one wanted to create multiple subvolumes for different purposes, the above procedure can be modified. For instance, if one wanted a separate subvolume mounted at / vs /home, then one can create two writable snapshots, empty out the contents of home in new subvolume 1 (but not the /home directory itself because you want the directory to exist for something to mount onto it), empty out everything outside of home within new subvolume 2, move the contents of home therein up one directory and remove the /home directory itself. Now, one can edit the fstab in new subvolume 1 as appropriate (not forgetting to have new subvolume 2 mount at /home), edit any other relevant files, reconfigure the bootloader to tell the Linux to use new subvolume 1 as the root subvolume, then reboot. Finally, one can remove the unnecessary files from the original root.
Edit:
It is arguably better to manually specify the new root when booting in the Linux kernel command line, and not reconfigure the bootloader until you successfully boot. After success, (if the following is relevant to your system) use update-grub, and it should look at fstab to automatically reconfigure the bootloader accordingly to use the appropriate new subvolume as specified at fstab.
This is what I did years ago to one of my own systems, although I don’t know anything about Timeshift and how it requires things to be set up (I have my own backup scripts that are run by cron). I could have just snapshotted the btrfs root directly for snapshots, but I wanted the snapshots to be cleanly separated from the subvolume used as the Linux VFS root (except when I explicitly mount them).
I had a case where fingerprint sensor was working out of the box fortunately. Although I had a problem where cryptfs would stop authenticating successfully with fingerprint sensor after distro update
Absolutely not outdated. I had a horrible time getting my hands on a working driver for the WiFi card in my brand new laptop last year. Horrible enough to resort to Ubuntu and even that gave me the finger. When I finally had it working I had to manually rebuild the damned thing each kernel update because I couldn’t convince DKMS to do it automatically. Had to wait two or three kernel releases for the card to be supported ‘out of the box’.
So no, fuck WiFI drivers in Linux. If it is not in the kernel and the manufacturer doesn’t provide one, don’t expect fun times.
Outdated for Linux Intel, still valid for Broadcom, probably not so bad for somewhat recent Realtek and AMD/Mediatek (last I’ve read is that Mediatek WiFi hardware sucks in general and disconnects happen on Windows, so the same happening on Linux would be the fault of the Linux driver).
I installed linux on a new pc 2 days ago, had no problem with the wifi drivers. I don’t know if it’s the fact that the wifi is integrated on the motherboard, but it was up and running without any tweeking from me (unlike windows)
But was the cause the Linux driver or the hardware? If the fault is the hardware and the experience on Linux is the same as on Windows, it’s feature parity.
If in doubt, get an Intel WiFi card. Even in otherwise not upgradeable notebooks those are usually not soldered on. Also whatever is in a Steam Deck OLED looks like a good pick.
Does Intel sell wifi cards that use USB rather than PCI slots? My motherboard doesn’t have the slot for a wifi PCIe card, and I’ve only seen Intel sell those :/
I can absolutely confirm it’s still valid for Realtek. I had one using the RTL8812AU chipset that basically no kernel version nor distro provided out of the box, so I constantly had to download a third-party driver from Github and manually patch it via dkms, or use a third-party repository containing the driver package… and then the driver broke so badly that it wouldn’t let me update at all unless I uninstalled it, which left me without the internet I needed to actually update, effectively leaving me unable to update until I could buy another one from Mediatek that’s compatible.
And said Mediatek wifi is really slow, so I just went from the frying pan into the fire…
I can absolutely confirm it’s still valid for Realtek. I had one using the RTL8812AU chipset
Yeah, and I was explicitly writing about recent chips. RTL8812AU isn’t recent. The very latest Windows driver is from 2018, so the chip itself was released a good while before that.
I know exactly what you had to go through because I had to do the same with mine a couple of years ago but since then for newer chips Realtek started contributing to Linux itself:
Ahh I see, thanks for clarifying. It seems that where I live mostly only has the older Realtek chips for sale, so I likely mostly had bad luck.
I tried USB tethering, but it wouldn’t work for some reason… I don’t remember exactly what happened, but I think either the phone or my computer couldn’t detect each other.
I do occasionally fall for just buying shtuff without a quick google search to see if my kernel would be cool with it, but I have an even greater number of stories about good experiences with Windows shtuff driving me bonkers.
For example, the Brother ADS-1200 under WIndows beats anything SANE supported scanners can do hands down. Scan to PDF with excellent compression and top of the line OCR. The spousal unit needed a scanner and I found a good deal on an ADS-2100. Under Linux, scan results are totally comparable to the ADS-1200, so the hardware is fine. But the Windows software for this scanner is crap. JPEG and TIFF are identical to the Linux scans, but OCR and PDF compression are atrocious. I’m 100% sure that if I were to edit a table in the ADS-1200 software, it would happily apply the same excellent results to the ADS-2100. But I’ve had it with hacking Windows goop, been there, done that, got the t-shirt, so onto Craig’s list the 2100 goes… Built in obsolescence, welcome to the Windows world.
With Linux, once the kernel accepts it, it’s smooth sailign without too many vendor introduced hickups.
And even on Windows, if you need to use third party scan software like VueScan because your scanner happens to be older than your Windows. it’ll work but it won’t outperform SANE supported scanners.
Situations like that aren’t very common these days. It usually happens when your hardware is very much new and drivers aren’t yet in the Linux kernel, or they are in the newest mainline, but your distro wont ship it for some more time. For that matter, it’s always bad when the kernel doesn’t have the drivers built in and it always requires dealing with DKMS or akmod whether it’s wifi, webcam, bluetooth or GPU (that’s why NVIDIA tends to be problematic on some systems).
That being said, the meme only works for anecdotal cases.
If it is not in the kernel and the manufacturer doesn’t provide one, don’t expect fun times.
This could be shorted to if your device has no driver it wont work which is obviously true.
If you have very recent hardware and you find it doesn’t work out of the box on stable options the easiest thing to do is install a more recent kernel. Even current Ubuntu non-LTS is 2-4 releases behind.
It’s even easier in arch/void where the latest kernel is already available.
Respectfully if DKMS wasn’t automatically kicking in then you configured it incorrectly. It’s a lot easier to just rely on a package that sets this up for you properly. If for some reason this can’t be done the logical thing to do is script the process so that all operations are completed in the appropriate order that way you needn’t remember to do one then the other.
This could be shorted to if your device has no driver it wont work which is obviously true.
What I tried to tell is that if you have to rely on community driver projects, don’t expect fun times, at least not when it comes to Realtek in my recent experience.
If you have very recent hardware and you find it doesn’t work out of the box on stable options the easiest thing to do is install a more recent kernel.
I already had the latest available kernel at the time, as in: the very latest officially released kernel by kernel.org. Ubuntu was just a last-ditch effort as it will sometimes have drivers included that other distros might not have, normally I wouldn’t touch it with a ten-feet pole and go either Arch or Manjaro. The driver simply wasn’t included in the kernel. How do I know? Because I stumbled upon some discussions that mentioned the lack of support and 3 kernel releases later support for my card was specifically mentioned in the changelog.
Respectfully if DKMS wasn’t automatically kicking in then you configured it incorrectly. It’s a lot easier to just rely on a package that sets this up for you properly.
Yes, like a Realtek-XXXX-dkms package, which simply didn’t work. I’ve configured stuff for DKMS before, scripting stuff for Linux is part of my daily workload, so yeah, you don’t need to tell me scripting beats doing stuff manually.
The fact that getting an f*cking wifi card to work takes this much effort is what I meant with ‘not fun times’ and for me validates the meme, anecdotal as it might be.
Resorting to other distros, configuring additional repos so you can install a different kernel version, having to try different community projects to see which gives you a working driver, having to deal with getting DKMS to work, this is all stuff which hampers Linux adoptment. And without more adoptment we won’t have to expect more support from manufacturers for desktop related consumer hardware. So yeah, that does make me cry a bit. It’s a catch-22 unfortunately.
dd does not stand for “disk duplicator”. That’s a modern backronymization that doesn’t reflect the original general usage of the command which is to “convert and copy”. Efficiently (with respect to I/O) copying raw data is only one of its intended purposes; it also converts text encodings.
I am gonna go against the circlejerk here and publicly admit that I have a macbook pro, daddy aint rich either, I wanted something that would last and works well, it’s both environmental and ux based decision for me, so it was either macbook or framework, since framework doesn’t sell in my country It’s an easy decision.
Also it’s an os most people want it to just work, one of the main reasons why iOS and MacOs are popular, until there is a linux flavor like mac is made and then it’s distributed as the default os, the year of the linux desktop will never come
You only have to use one of seemingly several methods to make the offline account option appear based on what specific version installer you have, no to Cortana but wait for her subtitles to finish in time with her audio anyway, individually untick every data collection option which each take up the whole screen with the toggle and next button being just far apart enough for it to be annoying plus the slow fade transition, realize you actually need the enterprise edition to set telemetry to 0 using group policy editor which isn’t available otherwise, have a vaguely different installation for that, find out that some functionality isn’t available like Ms store and some other stuff on enterprise which requires PowerShell to add it in if needed. Then possibly some random app to block select domains, with exceptions for the ones that make xbox, the Ms support sites, and ms software not work when disabled if needed. and/or pihole/unbound/Adguard if you have the means of setting it up. Then have random software not work for unknown reasons but you know deep down it’s the non standard installation of windows.
IKR, or pretty much any distro even some of the ‘advanced’ ones with some caveats.
I just had to do up a dual boot with windows for someone who works with people that use a windows only software that currently can’t be wine’d, and was appalled at how awful the install procedure has become since everything after 7 really. The enterprise and not pro features that are actually desirable for privacy minded people that aren’t experts in networking hardware and software was a huge let down too. Well assuming M$ isn’t just blatantly lying about config options anyway.
I remember having a conversation with a coworker who was getting into Linux when Kali was a big deal for script kiddies. He told me he installed it and I was like “dude you want it to be a read only OS, don’t install it. Just boot to it from a CD or USB.” We went back and forth on that for weeks until I just gave up and labeled him an idiot in my mind.
I put it on a dual boot laptop once because the laptop was to shitty to run to a proper VM and I wanted to get updates at a few different points in time. Intel Core 2 Dou and Windows XP as the other OS. It was more of a project laptop than a daily driver though.
Was windows XP the current windows generation or did you pick it for some other reason? I assume it’ll run easier on weak hardware, and until just now never thought about putting it on my laptop as a dual boot for those moments you actually need windows.
I’ve had this conversation with lots of first time Linux users. They think that Kali is the most hardcore hacker OS and that’s what they need to run for a introduction to security course.
When I was a kid I installed it and was like “hooHOO, me hacker”, so there are silly things like that.
Nevermind me being too intimidated by CLI to do anything in Linux at the time lmfao.
It’s been a while since I’ve thought about it, so what are the reasons why it’s a bad daily driver? I assume there’s poor support for drivers, hardware, etc.?
Or is it when you do pen testing you don’t want to leave traces of yourself? I’m not a cybersecurity guy, so I genuinely don’t know.
Hasn’t been the case anymore for quite some time, even though I think it has quite generous sudo rules. But yes, it’s not meant to be your main OS but instead more like a toolbox you use in liveboot/VM/etc.
linuxmemes
Active
This magazine is from a federated server and may be incomplete. Browse more on the original instance.