I usually interpret the phrase “drop in” to mean that the replacement being referenced will also work with everything written for the original. Does “drop in” in this case mean that Immich will transparently replace Google Photos, similar to how libretube replaces YouTube? That would be amazing!
It’s undergoing massive development, it basically went from nothing to nearly full featured in two years.
The breaking change just means you need to actually do something before updating. The software isn’t quite ready to be put on auto-update yet. Honestly the way the devs aren’t afraid to break things I think has contributed to the fast development.
Just be sure to keep a secondary backup of your photos which you should do either way.
I use it just as a simple as possible, instructions on how I setup backups, important thing about container’s config, etc etc. I find it easier to just have a folder “Server” and put each container in a separate note or folder. It’s too much thinking about tags, links, pages and all in logseq, notes seem all over the place.
Yep, now, I initially found the daily journal approach a bit strange, but I use this for work as much as personal stuff, so it actually helps…
My suggestion to your usecase would be to keep a page per “thing” ie server / container / etc and then when you make a change you can just say (on that day’s journal page):
‘’ Setup a backup for [[Server X]] and it’s going to [[NAS2]] (for example) ‘’
Then, on either of those 2 pages you’ll automatically see the link back to the journal page, so you’ll know when you did it…
I think you can disable the journal approach if it’s not useful…
But, the important part is, the files underlying the notes you’re making are in plain text with the page name as the filename, whereas with Joplin you could never find the file…
Also, if you modify the file (live) outside of Logseq, it copes with that and refreshes the content onscreen.
And the links are all dynamic… renamed the NAS? Fine, Logseq will reindex all the pages for you…
Seconding what others have already said. You should ABSOLUTELY NOT directly back up /var/lib/postgresql if that’s what you’re doing right now. Instead, use pg_dump: www.postgresql.org/docs/current/backup-dump.html
This should also give you smaller and probably more compressible backup sizes.
I would’ve NEVER ever moved to Joplin if it wasn’t able to sync with WebDAV. I’m not into having a special daemon running on a server for that task, makes zero sense.
Games from that time were actually running mostly in your browser. Meaning that the host, for example Miniclip served you the JavaScript and other files of the game which were then executed locally. So technically you could archive those games as long as you can load them up at least once initially.
If you logged and saved all the files the first one requested you could potentially make it work. You could manually change of the file paths in the html if you only doing a few of them. There’s only like 10 or so paths that would need to be modified. The PHP ones are likely harder to make work as php is a server side language and you don’t likely have easy access to PHP server and everything that goes with it.
Anyway thanks for the link to to mynoise.net. It looks like a well designed, carefully crafted website.
It’s an open source solution designed to scale to what the web was originally designed for and excels at. Documents. Specifically hyperlinked documents or webpages. You can’t reasonably expect an archival service to archive something that is by definition not static like an interactive web app.
It’s 2024, avoid Proxmox and safe yourself a LOT of headaches down the line.
You most likely don’t need Proxmox and its pseudo-open-source bullshit. My suggestion is to simply with with Debian 12 + LXD/LXC, it runs VMs and containers very well. Proxmox ships with an old kernel that is so mangled and twisted that they shouldn’t even be calling it a Linux kernel. Also their management daemons and other internal shenanigans will delay your boot and crash your systems under certain circumstances.
What I would suggest you to use use instead is LXD/Incus.
LXD/Incus provides a management and automation layer that really makes things work smoothly - essentially what Proxmox does but properly done. With Incus you can create clusters, download, manage and create OS images, run backups and restores, bootstrap things with cloud-init, move containers and VMs between servers (even live sometimes).
Another big advantage is the fact that it provides a unified experience to deal with both containers and VMs, no need to learn two different tools / APIs as the same commands and options will be used to manage both. Even profiles defining storage, network resources and other policies can be shared and applied across both containers and VMs.
I draw your attention to containers (not docker), LXC containers because for most people full virtualization isn’t even required. In a small homelab if you can have containers that behave like full operating systems (minus the kernel) including persistence, VMs might not be required. Either way LXD/Incus will allow for both and you can easily mix and match and use what you require for each use case.
For eg. I virtualize the official HomeAssistant image with LXD because we all know how hard is to get that thing running, however my NAS / Samba shares are just a LXD Debian 12 container with Samba4, Nginx and FileBrowser. Sames goes for torrent client that has its own container. Some other service I’ve exposed to the internet also runs a full VM for isolation.
Like Proxmox, LXD/Incus isn’t about replacing existing virtualization techniques such as QEMU, KVM and libvirt, it is about augmenting them so they become easier to manage at scale and overall more efficient. I can guarantee you that most people running Proxmox today it today will eventually move to Incus and never look back. It woks way better, true open-source, no bugs, no BS licenses and way less overhead.
I do regular snapshots of my containers live and sometimes restore them, no issues there. De-duplication and incremental features are (mostly) provided by the storage backend, if you use BTRFS or ZFS for your storage pool every container will be a volume that you can snapshot, rollback, export at any time. LXD also provides tools to make those operations: documentation.ubuntu.com/lxd/…/instances_backup/
create clusters, download, manage and create OS images, run backups and restores, bootstrap things with cloud-init, move containers and VMs between servers (even live sometimes).
provides a unified experience to deal with both containers and VMs, no need to learn two different tools / APIs as the same commands and options will be used to manage both. Even profiles defining storage, network resources and other policies can be shared and applied across both containers and VMs.
Your comment is wrong in a few ways and suggests using a LXC which is way slower than docker or podman and lacks the easy setup.
Proxmox is good because it makes it easy to create VMs and setup least access. It also has as new of kernel as stable Debian so no, its not terribly out of date.
If you want to suggest that someone install Debian + Docker compose that would make more sense. This isn’t a good setup for more advanced setups and it doesn’t allow for a not of flexibility.
This was a discussion about management solutions such as Proxmox and LXD and NOT about containerization technologies like Docker or LXC. Also Proxmox uses the Proxmox VE Kernel that is derived from Ubuntu.
Your comment makes no sense whatsoever. I’m not even sure you know the difference between LXD and LXC…
I’ve been on Proxmox for 6 or so months with very few issues and have found it to work well in my instance, I do appreciate seeing another alternative and learning about it too! I very specifically like Proxmox as it gives me an actual IP on my router’s subnet for my machines such as Home Assistant. So instead of the 192.168.122.1 it rolls a nice 192.168.1.X/24 IP which fits my range which makes it easier for me to direct my outside traffic to it. Does this also do this? Based on your screenshots, maybe not, IDK.
it gives me an actual IP on my router’s subnet for my machines
Yes you configure LXD/Incus’ networking to use a bridge and it will simply delegate the task to your router instead of proving IPs itself. One of my nodes actually runs the two setups at the same time, I’ve a bunch of containers on an internal range and then my Home Assistant VM getting an IP from my router.
Thanks for the link! I’ve been running Proxmox for years now without any of the issues like the previous commenter mentioned. Not that they don’t exist, just that I haven’t hit them. I really like Proxmox but love hearing about alternatives. One day I might get bored and want to set things up new with a different stack and anything that’s more free/open is better in my book.
Amazing! Have wanted something like this for years, currently use raindrop but not fully, very hesitant of locking myself in. This looks very promising.
I’m very curious… Why do you feel locked in by raindrop? I like that it can regularly upload exports to my Google drive and I can Always download them as html and csv.
selfhosted
Oldest
This magazine is from a federated server and may be incomplete. Browse more on the original instance.