209
submitted 2 months ago by data1701d@startrek.website to c/linux@lemmy.ml

I'm writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.

This has caused me to wonder what the largest storage operation you guys have done. I've taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.

top 50 comments
sorted by: hot top controversial new old
[-] Davel23@fedia.io 78 points 2 months ago

Not that big by today's standards, but I once downloaded the Windows 98 beta CD from a friend over dialup, 33.6k at best. Took about a week as I recall.

[-] pete_the_cat@lemmy.world 29 points 2 months ago

I remember downloading the scene on American Pie where Shannon Elizabeth strips naked over our 33.6 link and it took like an hour, at an amazing resolution of like 240p for a two minute clip 😂

[-] fahfahfahfah@lemmy.billiam.net 12 points 2 months ago

And then you busted after 15 seconds?

load more comments (1 replies)
[-] absGeekNZ@lemmy.nz 14 points 2 months ago

Yep, downloaded XP over 33.6k modem, but I'm in NZ so 33.6 was more advertising than reality, it took weeks.

load more comments (1 replies)
[-] Urist@lemmy.ml 56 points 2 months ago

I obviously downloaded a car after seeing that obnoxious anti-piracy ad.

[-] freijon@lemmings.world 56 points 2 months ago

I'm currently backing up my /dev folder to my unlimited cloud storage. The backup of the file /dev/random is running since two weeks.

[-] eager_eagle@lemmy.world 12 points 2 months ago

That's silly. You should compress it before uploading.

[-] Mike1576218@lemmy.ml 9 points 2 months ago

No wonder. That file is super slow to transfer for some reason. but wait till you get to /dev/urandom. That file hat TBs to transfer at whatever pipe you can throw at it...

[-] princessnorah 6 points 2 months ago

Cool, so I learned something new today. Don't run cat /dev/random

load more comments (2 replies)
[-] data1701d@startrek.website 5 points 2 months ago

I’m guessing this is a joke, right?

load more comments (2 replies)
[-] Neuromancer49@midwest.social 43 points 2 months ago

In grad school I worked with MRI data (hence the username). I had to upload ~500GB to our supercomputing cluster. Somewhere around 100,000 MRI images, and wrote 20 or so different machine learning algorithms to process them. All said and done, I ended up with about 2.5TB on the supercomputer. About 500MB ended up being useful and made it into my thesis.

Don't stay in school, kids.

[-] mEEGal@lemmy.world 17 points 2 months ago

You should have said no to math, it's a helluva drug

load more comments (2 replies)

Entire drive/array backups will probably be by far the largest file transfer anyone ever does. The biggest I've done was a measly 20TB over the internet which took forever.

Outside of that the largest "file" I've copied was just over 1TB which was a SQL file backup for our main databases at work.

[-] cbarrick@lemmy.world 8 points 2 months ago

+1

From an order of magnitude perspective, the max is terabytes. No "normal" users are dealing with petabytes. And if you are dealing with petabytes, you're not using some random poster's program from reddit.

For a concrete cap, I'd say 256 tebibytes...

load more comments (2 replies)
[-] Taleya@aussie.zone 36 points 2 months ago

I work in cinema content so hysterical laughter

[-] potajito@lemmy.dbzer0.com 14 points 2 months ago

Interesting! Could you give some numbers? And what do you use to move the files? If you can disclose obvs

[-] Taleya@aussie.zone 24 points 2 months ago* (last edited 2 months ago)

A small dcp is around 500gb. But that's like basic film shizz, 2d, 5.1 audio. For comparison, the 3D deadpool 2 teaser was 10gb.

Aspera's commonly used for transmission due to the way it multiplexes. It's the same protocolling behind Netflix and other streamers, although we don't have to worry about preloading chunks.

My laughter is mostly because we're transmitting to a couple thousand clients at once, so even with a small dcp thats around a PB dropped without blinking

[-] Azzk1kr@feddit.nl 11 points 2 months ago
[-] Dlayknee@lemmy.world 11 points 2 months ago

Digital Cinema Package; basically the movie file you're watching when you're in a movie theater.

load more comments (4 replies)
[-] MoonMelon@lemmy.ml 6 points 2 months ago

In the early 2000s I worked on an animated film. The studio was in the southern part of Orange County CA, and the final color grading / print (still not totally digital then) was done in LA. It was faster to courier a box of hard drives than to transfer electronically. We had to do it a bunch of times because of various notes/changes/fuck ups. Then the results got courier'd back because the director couldn't be bothered to travel for the fucking million dollars he was making.

load more comments (4 replies)
load more comments (3 replies)
[-] ramble81@lemm.ee 28 points 2 months ago

I’ve done a 1PB sync between a pair of 8-node SAN clusters as one was being physically moved since it’d be faster to seed the data and start a delta sync rather than try to do it all over a 10Gb pipe. M

[-] Hugin@lemmy.world 24 points 2 months ago

It was something around 40 TB X2 . We were doing a terrain analysis of the entire Earth. Every morning for 25 days I would install two fresh drives in the cluster doing the data crunching and migrate the filled drives to our file server rack.

The drives were about 80% full and our primary server was mirrored to two other 50 drive servers. At the end of the month the two servers were then shipped to customer locations.

[-] pixeltree 19 points 2 months ago

I once deleted an 800 gb log file, does that count

[-] Loulou@lemmy.mindoki.com 7 points 2 months ago

Depends, did you send it to the trash can first?

load more comments (1 replies)
[-] brygphilomena@lemmy.world 18 points 2 months ago

In the middle of something 200tb for my Plex server going from a 12 bay system to a 36 LFF system. But I've also literally driven servers across the desert because it was faster than trying to move data from one datacenter to another.

[-] Dlayknee@lemmy.world 9 points 2 months ago

That's some RFC 2549 logic, right there.

load more comments (3 replies)
load more comments (4 replies)
[-] Trigger2_2000@sh.itjust.works 17 points 2 months ago

I once abused an SMTP relay (my own) by emailing Novell a 400+ MB memory dump. Their FTP site kept timing out.

After all that, and them swearing they had to have it, the OS team said "Nope, we're not going to look at it". Guess how I feel about Novell after that?

This was in the mid-90's.

load more comments (1 replies)
[-] d00phy@lemmy.world 15 points 2 months ago

I’ve migrated petabytes from one GPFS file system to another. More than once, in fact. I’ve also migrated about 600TB of data from D3 tape format to 9940.

[-] HarriPotero@lemmy.world 11 points 2 months ago

I worked at a niche factory some 20 years ago. We had a tape robot with 8 tapes at some 200GB each. It'd do a full backup of everyone's home directories and mailboxes every week, and incremental backups nightly.

We'd keep the weekly backups on-site in a safe. Once a month I'd do a run to another plant one town over with a full backup.

I guess at most we'd need five tapes. If they still use it, and with modern tapes, it should scale nicely. Today's LTO-tapes are 18TB. Driving five tapes half an hour would give a nice bandwidth of 50GB/s. The bottleneck would be the write speed to tape at 400MB/s.

[-] MerchantsOfMisery@lemmy.ml 11 points 2 months ago

8 TB but I'm just a regular Joe with a penchant for piracy.

load more comments (2 replies)
[-] Decency8401@discuss.tchncs.de 11 points 2 months ago

A few years back I worked at a home. They organised the whole data structure but needed to move to another Providor. I and my colleagues moved roughly just about 15.4 TB. I don't know how long it took because honestly we didn't have much to do when the data was moving so we just used the downtime for some nerd time. Nerd time in the sense that we just started gaming and doing a mini LAN party with our Raspberry and banana pi's.

Surprisingly the data contained information of lots of long dead people which is quiet scary because it wasn't being deleted.

load more comments (1 replies)
[-] IphtashuFitz@lemmy.world 10 points 2 months ago

Back in the late 90’s I worked for an internet search company, long before Google was a thing. We would regularly physically drive a dozen SCSI drives from a RAID array between two datacenters about 20 miles apart.

[-] Larvitz@burningboard.net 10 points 2 months ago

@data1701d downloading forza horizon 5 on Steam with around 120gb is the largest web-download, I can remember. In LAN, I’ve migrated my old FreeBSD NAS to my new one, which was a roughly 35TB transfer over NFS.

load more comments (1 replies)
[-] avidamoeba@lemmy.ca 9 points 2 months ago* (last edited 2 months ago)

~15TB over the internet via 30Mbps uplink without any special considerations. Syncthing handled any and all network and power interruptions. I did a few power cable pulls myself.

load more comments (8 replies)
[-] weker01@sh.itjust.works 8 points 2 months ago

I once moved ~5TB of research data over the internet. It took days and unfortunately it also turned out that the data was junk :/

[-] NegativeLookBehind@lemmy.world 7 points 2 months ago

Approximately 2 petabytes.

[-] JerkyChew@lemmy.one 7 points 2 months ago

My Chia crypto farm at its peak had about 1.5 PB of plots, each plot was I think about 100ish gigs? I'd plot them on a dedicated machine and then move them to storage for farming. I think I'd move around 10TB per night.

It was done with a combination of powershell and bash scripts on Windows, Linux, and the built in Windows Services for Linux.

[-] delirious_owl@discuss.online 7 points 2 months ago

Upgraded a NAS for the office. It was reaching capacity, so we replaced it. Transfer was maybe 30 TB. Just used rsync. That local transfer was relatively fast. What took longer was for the NAS to replicate itself with its mirror located in a DC on the other side of the country.

load more comments (3 replies)
[-] MajorHavoc@programming.dev 7 points 2 months ago

I'll let you know... If it finishes.

[-] HappyTimeHarry@lemm.ee 7 points 2 months ago

I downloaded that 200gb leak from national public data the other day, maybe not the biggest total but certainly the largest single text file ive ever messed with

[-] psmgx@lemmy.world 6 points 2 months ago

Currently pushing about 3-5 TB of images to AI/ML scanning per day. Max we've seen through the system is about 8 TB.

Individual file? Probably 660 GB of backups before a migration at a previous job.

[-] southsamurai@sh.itjust.works 6 points 2 months ago

I think 16 terabytes? Might have been twelve. I was consolidating a bunch of old drives and data into a nas for a friend. He just didn't have the time, between working and school and brought me all the hardware and said "go" lol.

[-] lightnsfw@reddthat.com 5 points 2 months ago

4 TB over my home network. 800GB download from a external server.

[-] Yeahboiiii@lemm.ee 5 points 2 months ago

Largest one I ever did was around 4.something TB. New off-site backup server at a friends place. Took me 4 months due to data limits and an upload speed that maxed out at 3MB/s.

[-] iopq@lemmy.world 5 points 2 months ago

I synced to the BSV shitcoin which is 11+ terabytes. So large I had to turn on throwing away the rest of what I downloaded because it wouldn't fit on all of the storage media I own. I feel sorry for the people running an archive node.

[-] Joelk111@lemmy.world 5 points 2 months ago* (last edited 2 months ago)

When I was moving from a Windows NAS (God, fuck windows and its permissions management) on an old laptop to a Linux NAS I had to copy about 10TB from some drives to some other drives so I could re-format the drives as a Linux friendly format, then copy the data back to the original drives.

I was also doing all of this via terminal, so I had to learn how to copy in the background, then write a script to check and display the progress every few seconds. I'm shocked I didn't loose any data to be completely honest. Doing shit like that makes me marvel at modern GUIs.

Took about 3 days in copying files alone. When combined with all the other NAS setup stuff, ended up taking me about a week just in waiting for stuff to happen.

I cannot reiterate enough how fucking difficult it was to set up the Windows NAS vs the Ubuntu Server NAS. I had constant issues with permissions on the Windows NAS. I've had about 1 issue in 4 months on the Linux NAS, and it was much more easily solved.

The reason the laptop wasn't a Linux NAS is due to my existing Plex server instance. It's always been on Windows and I haven't yet had a chance to try to migrate it to Linux. Some day I'll get around to it, but if it ain't broke... Now the laptop is just a dedicated Plex server and serves files from the NAS instead of local. It has much better hardware than my NAS, otherwise the NAS would be the Plex server.

load more comments (1 replies)
[-] Someonelol@lemmy.dbzer0.com 5 points 2 months ago

Manually transferred about 7TBs to my new Rpi4 powered NAS. It took a couple of days because I was lazy and transferred 15 GBs at a time which slowed down the speed for some reason. It could handle small sub 1 GB files in half a minute otherwise.

[-] hperrin@lemmy.world 5 points 2 months ago

I transferred my entire NAS storage, which includes all of my backups, cloud files, my family’s backups, and my… Linux ISOs. That was about 12TB.

[-] nik9000@programming.dev 5 points 2 months ago

When I was in highschool we toured the local EPA office. They had the most data I've ever seen accessible in person. Im going to guess how much.

It was a dome with a robot arm that spun around and grabbed tapes. It was 2000 so I'm guessing 100gb per tape. But my memory on the shape of the tapes isn't good.

Looks like tapes were four inches tall. Let's found up to six inches for housing and easier math. The dome was taller than me. Let's go with 14 shelves.

Let's guess a six foot shelf diameter. So, like 20 feet circumference. Tapes were maybe .8 inches a pop. With space between for robot fingers and stuff, let's guess 240 tapes per shelf.

That comes out to about 300 terabytes. Oh. That isn't that much these days. I mean, it's a lot. But these days you could easily get that in spinning disks. No robot arm seek time. But with modern hardware it'd be 60 petabytes.

I'm not sure how you'd transfer it these days. A truck, presumably. But you'd probably want to transfer a copy rather than disassemble it. That sounds slow too.

load more comments (3 replies)
load more comments
view more: next ›
this post was submitted on 18 Aug 2024
209 points (100.0% liked)

Linux

47819 readers
525 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS