Thread refresh

Xe :verified: @cadey@pony.social
replies: #2 #5 #9 #14 #16 #17

SRE graph scrying test:

I'm backing up a bunch of files to S3, there's moments where the throughput goes down and moments where it goes up.

Assuming no network activity in a way that matters, why would the throughput number be going up and down like this?

Daniel L @glitchf@octodon.social
in reply to #1 - replies: #3

@cadey The numbers going down mean lots of little small files. Going up means big long transfers.

Cal Paterson @calpaterson@fosstodon.org

@cadey @glitchf some of the better s3 tooling (eg rclone) have ways to mitigate this via more concurrency

luciano @luciano@parens.social
in reply to #1 - replies: #6

@cadey disk go brr vs disk go 🥶?

Xe :verified: @cadey@pony.social
in reply to #5 - replies: #7

@luciano What makes you think that?

Mikle_Bond @Mikle_Bond@pony.social
in reply to #6 - replies: #8

@cadey @luciano
on a side note, it might be close being correct. A drive can show this access pattern at some point of us life, usually before death.

Here's your unscheduled reminder to check if smartd is configured properly)

Xe :verified: @cadey@pony.social

@Mikle_Bond @luciano I mean I'm also doing a ZFS resilvering because a drive died lol

Xe :verified: @cadey@pony.social

@Schneems Nope! The remote service (which is like S3 but not S3) is able to accept data at my NAS' line rate, and my drives are able to supply data also at line rate.

Richard Schneeman @Schneems@ruby.social

@cadey super weird.

Xe :verified: @cadey@pony.social
Spoilers

@Schneems Biblical amounts of small files, there's more overhead in creating lots of little files than there are creating very few big files

Mikle_Bond @Mikle_Bond@pony.social
Spoilers

@cadey @Schneems
The difference in throughput becomes even more drastic with SMB on windows and even more educational.

fogs :pride: @andrew@montagne.uk

@cadey lots of small files vs small number of big files?

@lillian

@cadey a LLM is generating `aws s3 cp` commands for each file and different filenames (presumably that parse into more tokens) take longer for the model to generate commands for

AMS @AMS@infosec.exchange

@cadey Copying kernel source or some similar pile of tiny files?