Linux desktop will, most likely, fail for: Developers and sysadmins, because not everyone is using Docker and Github actions to deploy applications to some proprietary cloud solution. Finding a properly working FTP/SFTP/FTPS desktop client (similar WinSCP or Cyberduck) is an impossible task as the ones that exist fail even at basic tasks like dragging and dropping a file.
Linux desktop is great, I love it but I don’t sugar coat it nor I’m delusional like most posting about it.
It all comes down to a question of how much time (days? months?) you want to spend fixing things on Linux that simply work out of the box under Windows for a minimal fee. Buy a Windows license and spend the time you would’ve spent dealing with Linux issues doing your actual job and you’ll, most likely, get a better ROI.
Guess what happens whenever people popularize immutable distros as the next hype in tech that will make everything better? You get yourself into a totally unreasonable and avoidable ecosystem just because those systems won’t cut it for most use cases… same that happened with Docker/Kubernetes.
I’ve been saying it for year and nobody cares: nowadays those companies are all about re-creating and reconfiguring the way people develop software so everyone will be hostage of their platforms. We see this in everything now Docker/DockerHub/Kubernetes and GitHub actions were the first sign of this cancer. We now have a generation of developers that doesn’t understand the basic of their tech stack, about networking, about DNS, about how to deploy a simple thing into a server that doesn’t use some Docker BS or isn’t a 3rd party cloud xyz deploy-from-github service.
The latest endeavor in making everyone’s hostage is the new Linux immutable distribution trend. Immutable distros are all about making thing that were easy into complex, “locked down”, “inflexible”, bullshit to justify jobs and payed tech stacks and a soon to be released property solution.
We had Ansible, containers, ZFS and BTRFS that provided all the required immutability needed already but someone decided that is is time to transform proven development techniques in the hopes of eventually selling some orchestration and/or other proprietary repository / platform / BS like Docker / Kubernetes does.
“Oh but there are truly open-source immutable distros” … true, but this hype is much like Docker and it will invariably and inevitably lead people down a path that will then require some proprietary solution or dependency somewhere that is only required because the “new” technology itself alone doesn’t deliver as others did in the past.
As with CentOS’s fiasco or Docker it doesn’t really matter if there are truly open-source and open ecosystems of immutable distributions because in the end people/companies will pick the proprietary / closed option just because “it’s easier to use” or some other specific thing that will be good on the short term and very bad on the long term. This happened with CentOS vs Debian is currently unfolding with Docker vs LXC/RKT and will happen with Ubuntu vs Debian for all those who moved from CentOS to Ubuntu.
Those popularizing immutable distributions clearly haven’t had any experience with it before the current hype. Let me tell you something, immutable systems aren’t a new thing we already had it with MIPS devices (mostly routers and IOTs) and people have been moving to ARM and mutable solutions because it’s better, easier and more reliable.
There is always some solutionizm in tech, but I’m interested in containerzation as a solution to problems I’ve had with configure drift building up on my systems and make it easier to share and work with the community.
The immutable desktop work to me is specifically working on bridging the gap between the UX of a local admin (you know wanting custom configuration and fast reaction to user input) and the industrial expectations of being able to test and track every change and reduce the number of different pieces you need to operate a system.
Hopefully we can lose some of the industries bad habits though. Like “relying on this proprietary piece is ok because we can move faster” or making other excuses as if you are going to have to explain to your boss why some metric looks bad instead of just trying to make the best system or solution we can.
The amount of times my Windows installation(s) broke is just as high as the amount of times my Linux installations had issues. The article you quoted seems to be from someone with more Windows experience than Linux experience.
One example: FileZilla is a capable GUI SFTP and FTP client, but so is nearly every file manager. I can drag and drop files from Dolphin into a fuse mounted FTP, SMB or SFTP folder just fine. Skill issue?
EDIT: omg, I just realised they use WinSCP for deploying applications. It really seems like a skill issue since you can automate that even without proprietary clouds. I can probably replace this person with a PowerShell script, which is even more efficient than them doing their job on Windows.
All operating systems sadly need lots of maintenance nowadays. The main reason I use Linux is that I feel in control of the system and the vendor doesn’t actively try to fuck with my installation.
Idk for the how but airvpn does for comparable prices. This coming from a fellow multi-tb Linux iso torrenter. Also I assume you mean VPN unless mullvad does VPS stuff I don’t know of.
Does your network not support UPnP? You shouldn’t normally need to port forward in order to seed a torrent, unless your network prevents NAT traversal.
This project is currently in a very early stage of development. Kando is not yet a functional menu but rather a prototype which demonstrates the feasibility of the concept.
Since Kando is still in early development, it might be a good idea to look at the Gnome Extension Fly-Pie. It’s from the same developer and it looks like Kando will be similar.
Near double the size and resources for having a dock/bar/menu and pinning icons on the background .. too much clutter for things hiding behind whatever you are doing most of the time.
A desktop is something you use to impress someone using mac/msWin ...
Not really relevant, but I’ve got a “rule of thumb” for all security-related issues;
“If it doesn’t nuke my PC, then I’m good. If it does, then I’m still good since backups and logs exist, and if it was related to the latest seucirty issue? Then I make a quick patch and/or update. Then back to 1.”
"Cloud native" technology is double speak for your shit is running on other people's computers who will be tracking your use and selling it to pay for server upkeep and also maybe profit?
In this case it's referring to the fact that the OS is built upon the same containerization technology used on cloud platforms such as Kubernetes. As a marketing tool it's a bit buzzwordy, but it's not about running the core OS components outside of the physical machine here.
Both are incredibly stupid attempts to convince people they need something they don’t.
You all should look into a ‘blue ocean’ business strategy. Lots of shitty businessmen are constantly trying to push ‘new’ things and have a vested interest in convincing laymen of their necessity.
Needs are born from solving problems, not making them.
I think you can use Arch for university, but I have a few suggestions:
Don’t update packages unless you have time to fix issues. Rare, but it happens, although usually minor.
Never mess with greeters, kernel modifications, bootloaders, or anything else before login. Fixing issues may require a live USB and take some time. Avoid the temptation! (Ask how I know…)
Use Flatpak where you can for increased stability by way of fewer packages to update.
Of course, you could also use a non-rolling release distro. Nothing wrong with that.
Excuse me if this is a bit of a dumb question, as I have never particularly worried about packaging methods and simply installed what I needed from the official Arch packages or AUR, but how does Flatpak lead to fewer updates? I know it sandboxes things, that’s why I’ve been interested in it for applications I don’t quite trust like Discord, but I never got around to actually switching applications of that sort over and trying the format out.
Speaking of Discord, hooking that out of the “normal packages”, aka everything I update via yay, would be beneficial anyways, since it’s the only thing that forces me to update my system by saying how I’m oh so lucky about a new update coming out and I don’t wanna mess with partial system updates. That’s kinda besides the point though, I just wanted to complain.
Good question. Flatpak doesn’t lead to fewer updates overall, but it does lead to fewer system packages installed via pacman or yay, which can run into dependency conflicts unlike Flatpak.
Flatpak provides a common runtime upon which different applications can be installed via containers, much like apps on a phone. You can then adjust the permissions for each app such as which directories it can access. It’s kind of like installing Firefox (e.g. Flatpak) and then a Firefox extension such as uBlock Origin (e.g. LibreOffice). It doesn’t matter if you’re on Ubuntu, Fedora, or Arch, from the extension’s perspective it’s the same old Firefox and doesn’t touch anything else on your computer. This means uninstalling is clean and it has no dependencies on other packages installed.
The disadvantage with Flatpak is they can be slower (sometimes not even noticeable) and take up more space, although the runtime can be re-used between Flatpak applications. Personally, I like Flatpak for large GUI applications like LibreOffice, which has 170+ dependencies if installed via pacman.
Huh, neat. On my current EOS Desktop PC it’s a bit too late to uninstall everything and reinstall it as a Flatpak, but I’ll keep that in mind for new Linux installs. Thank you :3
linux
Newest
This magazine is from a federated server and may be incomplete. Browse more on the original instance.