Yea I’ve been using nextcloud for a while and it’s fine.
I remember when I used owncloud before nextcloud was even a thing and the upgrade experience was absolute shit.
These days it’s just fine.
I disagree–a system (even Arch!) should be able to update after a couple months and not break! I recently booted an EndeavourOS image after 6 months and was able to update it properly, although I needed to completely rebuild the keyring first
Arch and EndeavourOS are the same thing. There is no functional difference between using one or the other. They both use pacman and have the same repos.
Very true–the specific EOS repo has given me a bit of trouble in the past, but it takes like 3 commands to remove it and then you’ve got just arch (although some purests may disagree 🤣)
I know this is how it’s supposed to be and how it should be but sadly it doesn’t always go this way and arch is notoriously known for this exact problem, the wiki itself tells you to check what’s being upgrades before doing because it might break. Arch is not stable if you don’t expect it to be unstable.
I’m using opensuse tumbleweed a lot - this summer I’ve found an installation not touched for 2 years. Was about to reinstall when I decided to give updating it a try. I needed to manually force in a few packages related to zypper, and make choices for conflicts in a bit over 20 packages - but much to my surprise the rest went smoothly.
I regularly “deep freeze” or make read-only systems from Raspberry Pi, Ubuntu, Linux Mint LMDE and others Linux Distros whereas I disable automatic updates everywhere (except for some obvious config/network/hardware/subsystem changes I control separately).
I have had systems running 24/7 (no internet, WiFi) for 2-3 years before I got around to update/upgrade them. Almost never had an issue. I always expected some serious issues but the Linux package management and upgrade system is surprisingly robust. Obviously, I don’t install new software on a old system before updating/upgrading (learned that early on empirically).
Automatic updates are generally beneficial and helps avoid future compatibility/dependency issues on active systems with frequent user interaction.
However, on embedded/single purpose/long distance/dedicated or ephemeral application, (unsupervised) automatic updates may break how the custom/main software may interact with the platform. Causing irreversible issues with the purpose it was built for or negatively impact other parts of closed circuit systems (for example: longitudinal environmental monitoring, fauna and flora observation studies, climate monitoring stations, etc.)
Generally, any kind of update imply some level of supervision and testing, otherwise things could break silently without anyone noticing. Until a critical situation arises and everything break loose and it is too late/too demanding/too costly to try to fix or recover within a impossibly short window of time.
Dude- it’s like you’re reading my mind. I’ve installed Nextcloud 4 different times, the most recent being on docker desktop in Win11. I’ve resorted to using chatgpt to help me with the commands. LITERALLY EVERY STEP RESULTS IN AN ERROR. The Collabora office suite (necessary to view or edit cloud docs without downloading them) WILL NOT DOWNLOAD. The “php -d memory_limit=512M occ app:install richdocumentscode” chatgpt and Nextcloud suggest is not recognized by the terminal. You can’t just download Collabora, cuz fuck you, i guess, and you can’t access Docker’s actual file system from windows explorer.
I’ve typed nonsense into various black screens for upward of 20 hours now, and nextcloud is “working” locally. I can access my giant hard drive from my android nextcloud app, but it’s SLOW AS FUCK.
I can’t imagine how many man-hours it would take to open the server to the internet. Makes me want to fucking barf just thinking about it.
I’ve been fucking with Linux since 2005 and have yet to get a single thing to work correctly. I guess I’m the only one who thinks an (mostly) invisible file system in incomprehensible repetitive folders, made of complete nonsense commands might not be the best way to operate a computer system.
I’m really frustrated if you can’t tell.
On another topic, trying to get Ollama to run on my Lubuntu VM was also impossible. I guess if everyone knew it was going to force you to somehow retroactively configure every motherfucking aspect of the install nobody would bother. You can sudo all day and it still denies me permission to do things LISTED IN THE MOTHERFUCKING DOCUMENTATION.
Is this all just low-effort poorf** bullshit that doesn’t actually work?
The problem child for me right now is a game built in node.js that I’m trying to host/fix. It’s lagging at random with very little reason, crashing in new and interesting ways every day, and resisting almost all attempts at instrumentation & debugging. To the point most things in DevTools just lock it up full stop. And it’s not compatible with most APMs because most of the traffic occurs over websockets. (I had Datadog working, but all it was saying was most of the CPU time is being spent on garbage collection at the time things go wonky–couldn’t get it narrowed down, and I’ve tried many different GC settings that ultimately didn’t help)
I haven’t had any major problems with Nextcloud lately, despite the fragile way in which I’ve installed it at work (Nextcloud and MariaDB both in Kubernetes). It occasionally gets stuck in maintenance mode after an update, because I’m not giving it enough time to run the update and it restarts the container and I haven’t given enough thought to what it’d take to increase that time. That’s about it. Early on I did have a little trouble maintaining it because of some problems with the storage, or the database container deciding to start over and wipe the volume, but nothing my backups couldn’t handle.
I have a hell of a time getting the email to stay working, but that’s not necessarily a Nextcloud problem, that’s a Microsoft being weird about email problem (according to them it is time to let go of ancient apps that cannot handle oauth2–Nextcloud emailer doesn’t support this, same with several other applications we’re running, so we have to do some weird email proxy stuff)
I am not surprised to hear some of the stories in this thread, though. Nextcloud’s doing a lot of stuff. Lots of failure points.
I’ve just finally and fully spun down a proxmox server I’ve been running and updating as my home lab for six years.
Every major update seemed to break something. Upgrades were always a roll of the dice as to whether it would even boot. It’s probably at least partially my fault for using an old R710 and running docker directly on the OS instead of within a container, but it was still by far my least reliable piece of kit.
The last apt update removed sudo, and I can’t be arsed to rebuild, so I’ve moved the critical bits to a fleet of SBCs. Powering that fucker down was a huge relief.
The new Linuxserver.io docker image at the very least has solved the annoying update cycle NextCloud has and seems to have fixed the need to do that every few months. I haven’t ever had it die but I don’t push it hard and I keep the plugins to a minimum because I just don’t trust it and it doesn’t run all that well.
For years, I had an unstable unraid server. I was fixing it every couple of days after a lockup. I had decided that unraid sucked. When it was up for a week I celebrated. Every one of my dockers was a suspect. I learned to hate all of them.
Never had a single functional problem with Nextcloud, other than the fact that it’s oppressively slow with the amount of files I’ve shoved into it. Mind you I also don’t use MySQL/MariaDB which I consider a garbage-tier DB. Despite Postgres not being the “Recommended DB” for Nextcloud it works perfectly for me. Maybe that’s the difference.
Nextcloud for me too, would break because of updates requiring manual DB updates sometimes, apps would randomly stop working after updating too, or the 2 times it caused total data loss on all my synced devices and the server itself which required a full restore from backups.
After getting rid of it and switching to Syncthing + Filebrowser + SFTPGo for WebDAV I haven’t really had anything break since then (about a year now). Stuff also runs much faster, NC was extremely slow even on good hardware with all their recommended settings for performance.
If Nextcloud “caused total data loss on all my synced devices and the server itself” I would probably do something unsavory to any responsible party I could locate, and take 10 TB of data out of their lousy hide.
Yeah the first time was the time/date bug they had (still have?) where it set the time on every folder and file to 00/00/0000 00:00 across all clients and the server.
Second time was I disabled virtual file support on my laptop so it would sync everything, but instead it went and wiped all the files from the server, because for some reason their sync client assumed the laptop that now had no files on it should be the master source or something.
Their own docs even state that’s how you’re supposed to disable VFS, with no mention that it will wipe your server clean.
Yeah the Docker version hated me, mainly due to it sometimes getting a bit behind on updates and then having schema mismatches if I ran an update in that missed the previous one. No issues with the Snap thus far
I used to have this problem. I started pulling a version number (like 27) instead of “latest” so that I could just pull minor releases when I did updates, and then I manually step up the version in the docker-config file for major versions when I’m ready for them. (I don’t like to pull a major release version until there’s been 1 or 2 maintenance releases since my nextcloud is fairly critical for my family)
The solution for me is that I run Nextcloud on a Kubernetes cluster and pin a container version. Then every few months I update that version in my deployment yaml to the latest one I want to run, and run kubectl apply -f nextcloud.yml and it just does its thing. Never given me any real trouble.
Well… no… I have been self hosting it for several years over multiple major versions now. Only for Files, Calendar and Deck though. It was a bit hard to set up, but reading the general Apache and PHP documentation helped a lot.
I’ve setup Nextcloud but have done next to nothing with it.
My Lemmy instance gives me the most problems, but it’s also the only publicly available service I run. Mostly the issue is it seems to have a memory leak that forces me to restart it every few days.
Everything else has been completely rock solid for me, running on a mini pc (formerly a pi4 until I wanted to start doing stuff with Jellyfin and needed more power for transcoding) on OpenSUSE Leap all in docker containers. Makes it insanely easy to move stuff. I had no issues basically just copying the docker-compose files and data and bringing them up even when switching architectures.
Add comment