datahoarder

This magazine is from a federated server and may be incomplete. Browse more on the original instance.

Atemu, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)
@Atemu@lemmy.ml avatar

I don’t want to do any sort of RAID 0 or striping because the hard drives are old and I don’t want a single one of them failing to make the entire backup unrecoverable.

This will happen in any case unless you had enough capacity for redundancy.

What is in this 4TB drive? A Linux installation? A bunch of user data? Both? What kind of data?

The first step to this is to separate your concerns. If you had e.g. a 20GiB Linux install, 10GiB of loose home files, 1TiB of Movies, 500GiB of photos, 1TiB of games and 500GiB of Music for example, you could back each of those up separately onto separate drives.

Now, it’s likely that you’d still have more data of one category than what fits on your largest external drive (movies are a likely candidate).

For this purpose, I use git-annex.branchable.com. It’s a beast to get into and set up properly with plenty of footguns attached but it was designed to solve issues like this elegantly.
One of the most important things it does is separate file content from file metadata; making metadata available in all locations (“repos”) while data can be present in only a subset, thereby achieving distributed storage. I.e. you could have 4TiB of file contents distributed over a bunch of 500GiB drives but in each one of those repos you’d have the full file tree available (metadata of all files + content of present files) allowing you to manage your files in any place without having all the contents present (or even any). It’s quite magical.

Once configured properly, you can simply attach a drive, clone the git repo onto it and then run a git annex sync --content and it’ll fill that drive up with as much content as it can or until each “file”'s numcopies or other configured constraints are reached.

the_q, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)

This seems like a terrible idea.

keefshape, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)

Mergerfs to combine the smaller ones.

athos77, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)

What I do is on the originator drive, I create new subdirectories and start categorizing items by content; like I'll put all the ebooks into one directory, and all the television into another. It just makes it easier for me to find things later if I can just head to the drive with all the television on it.

If there's a particular directory with a lot of content, I might create further divisions - maybe shows that are finished vs those who are still getting new episodes, or sitcoms vs drama, that kind of thing.

Then I make a list of how big each master directory is, and I start copying them over to the most appropriate-sized drive. I usually find that I can fit in one large directory, and a couple of smaller ones, and then the last drive gets all the leftovers. I also tape a post-it note to each drive saying something like "2022-23 television" or "science fiction audiobooks" or whatever.

I also create a new directory on the originating drive called something like ++COPIED and, once I've copied content to a new drive, I move the original directory to ++COPIED: I'll still have access if I need it, but I don't have to keep track of it any longer. Once everything is successfully copied over, I can just delete that one directory.

It's a manual process, yes, but it does make it easier for me to find stuff when I want to look at it again later.

mindlessLump, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)

You could create a Python script to do this. There is a library called psutil that would help. Basically,

  • iterate over mounted drives and see how much each has available
  • based on these values, iterate over your backup files and separate them into chunks that will fit on each drive
  • copy chunks to respective drives

Would be a fun little project even for a beginner I think.

cerement, in What's an elegant way of automatically backing up the contents of a large drive to multiple smaller drives that add up to the capacity of the large drive? (on Linux)
@cerement@slrpnk.net avatar

sounds like your main limitation is attaching the drives – if you can attach them all to a single system (ex. a separate computer or a NAS case) then at least it becomes somewhat easier to access them all at once

I was thinking JBOD but Wikipedia points out the same issue you mention with RAID 0, failure of one drive can mess up the logical volume which leads to a whole host of new issues to deal with during recovery

ColdCreasent, (edited )

Not that big of a deal when it is a backup. Raid is not a backup solution, it is a 24/7 uptime solution. If the main drive dies with JBOD, then you have the backup. If a backup drive fails, then you still have the main. Trick is to ensure any drive issues are dealt with immediately and no backup runs if there is a smart error or similar on any drive.

So having software that monitors drive health and email/notifies you is necessary.

Secondary benefit of JBOD is all drives in a pool are still readable separately.

SplicedBrainwrap, in How do I archive a podcast I've already downloaded in Playapod (iOS) that has since gone offline?

The Playapod website mentions a web player, does the web player happen to sync with your app on the phone? If so maybe you can extract it through the web player on a desktop.

yo_scottie_oh,

Thanks for the idea. I figured out how to get what I wanted and edited the solution into my post. Thanks again!

bela, (edited ) in How do I archive a podcast I've already downloaded in Playapod (iOS) that has since gone offline?

According to this you can copy stuff onto it. I assume you’ve tried to see if you can copy off? If not, then there is probably no way to extract the audio, short of recording it as it plays.

I was only able to find a couple episodes online. One on someone’s google drive and a few in the wayback machine, all from this thread

edit: if playapod has an option to store files on an sd card, that’s usually stored unencrypted. (on android anyway…)

edit: wait iphones don’t even have sd card slots, right? lol

bela, (edited )

Actually there might be a whole bunch more in the wayback machine. (though still not that many) here

Just sort by mime type and check out audio/* for items of interest

yo_scottie_oh, (edited )

This helped me, thank you so much! I filtered by .mp3 and I’m using DownThemAll to download all the episodes in batches.

ETA at first, all the downloads failed because I was attempting to download multiple files simultaneously. After configuring DownThemAll to allow only one simultaneous download, it’s been smooth sailing.

watson387, (edited ) in What to do with extra HDDs
@watson387@sopuli.xyz avatar

WTF is an extra HDD?

netburnr, in What to do with extra HDDs
@netburnr@lemmy.world avatar

A mirror array is not a backup. So therefor I would use at least one of those extra drives for a weekly backup of your data. You want some sort of not real-time backup in case you get cryptoed for example

jh0wlett, (edited )

That is true, but I was indeed not counting it as such, I currently just have the cloud as a backup.

netburnr,
@netburnr@lemmy.world avatar

If you’ve got cloud backup you’re good, keep the drives in a drawer as spares or have another use?

Efwis, in What to do with extra HDDs

Primary rule of thumb is 2 physical backups and one offsite backup(cloud). Remember physical medium can fail or degrade

jh0wlett,

I understand, but I was thinking, since I have a cloud backup, if my drive(s) fail I cam still always recover, correct?

Efwis,

What if the cloud server corrupts your data in transfer or worse shuts down its server without notification. It can and has happened.

For example, I had a cloud backup went to get it and the server could no longer be found. That was with Dropbox mind you. I lost 10gb of important files because of it. Never trust just one source of backup. Always have a secondary just in case.

Owljfien,

Cloud service may also decide that they’re tired of being in business and close without notice. Unlikely, but possible

Reverendender, in What to do with extra HDDs

Is it ok to ask what sort of data you have?

jh0wlett,

Yes definitely! It is personal data like family videos, photo’s and documents mostly. I’m using this nextcloud for the whole family.

pbjamm, in A dozen or two TB of storage for media on the cheap?
@pbjamm@beehaw.org avatar

Check out the Asustor AS5202T

You could get it and a pair of 8TB HDDs for around your $500 mark. Or spring for the 4 bay model and get cheaper drives.

Lemmchen, (edited ) in A dozen or two TB of storage for media on the cheap?

Unfortunately I’m not familiar with the US market, but in Europe we have sites like Geizhals (“Skinflint” in the UK) that are excellent at listing electronics, so you could source them for cheap: internal drives, external drives

If whatever you end up doesn’t have enough 3.5" slots, you can get these things on ebay to use them as a hacked together external housing:
https://feddit.de/pictrs/image/d34e7bd0-5f78-4e3c-a5b7-56430e8482fb.png
You’d need to get the SATA cables out of your case though.

As for what drives to use: If you don’t need redundancy/parity, then a single 12 (14, 16, 18, whatever you need) TB SATA drive will probably beat everything else pricewise. I’d say that leaves you with roughly $300 for the system itself, if you need to buy a new one.

but doesn’t have any way of adding a bunch of drives

Well, you only really need one or two drives. Are you sure it doesn’t offer any SATA connections?
What about PCIe? You could use a cheap HBA card then.

digdilem, in A dozen or two TB of storage for media on the cheap?

I’m sure plenty will disagree with me, but unless you have specific needs, I’d suggest spending more time sourcing your media rather than rely on transcoding. Most formats of popular stuff are available and Jellyfin will happily play it natively.

Also be aware that transcoding is VERY cpu intensive, unless you have a compatible gpu/transcoder. I run a ML110 with a 6-core Xeon (12 threads) and if Jellyfin needs to transcode something, it uses all of that and still stutters badly when seeking.

If you do need to transcode because you can’t source the media in a compatible way - you may want to use something like Tdarr to transcode it before you try to play it, so it’s ready when you are.

Lemmchen, (edited )

I’m running a Ryzen 7 2700X without any GPU acceleration and I can transcode any media just fine with Jellyfin.

beerclue,

An i7 gen 10 with QS can handle transcoding way better than a Xeon…

  • All
  • Subscribed
  • Moderated
  • Favorites
  • datahoarder@lemmy.ml
  • localhost
  • All magazines
  • Loading…
    Loading the web debug toolbar…
    Attempt #