@Lord is a user on octodon.social. You can follow them or interact with them if you have an account anywhere in the fediverse. If you don't, you can sign up here.

Anyone have any tips for working with large text files? (~30Gb)

I'm assuming that pretty much any normal text editor is out of the question? Is a combination of sed and grep my best bet?

This almost seems like a case for Ed, man!
gnu.org/fun/jokes/ed-msg.html

@codesections Use grep to separate them into separate files by first character?

Use /usr/bin/split to section the data into separate files?

@Lord @lord

@drwho @codesections This is the way to go.

· Web · 0 · 0

@lord @drwho Yeah, that does seem like the most promising option. I'd thought of splitting the file, but splitting based on letter is much better than just arbitrarily. Thanks, both!

@codesections @drwho Note that you could virtually store it virtually without any storage.
Creating folders with only one letters over and over until you have the complete hash.

0 space used in the filesystem !
You won't need ram anymore too !

:blobamused:

It reminds me this story : patrickcraig.co.uk/other/compr

@lord @drwho (also, I've seen that story before. It's a good one!)

@lord @codesections That's a really interesting idea... wouldn't you eventually run out of inodes, though?

@drwho @codesections It was a joke.
It's doable, you won't run out of inodes with modern filesystem (btrfs can have 2^64 files/folders) but it won't magically give you free space.

Your files will weight 0 but your filesystem metadata will be huge…

@lord @drwho Hence why the story @lord linked is so on-point—it also relied on "hiding" the size of data in the meta-data