Beefy Boxes and Bandwidth Generously Provided by pair Networks
Perl-Sensitive Sunglasses
 
PerlMonks  

Re^2: 15 billion row text file and row deletes - Best Practice?

by brig (Scribe)
on Dec 01, 2006 at 20:09 UTC ( [id://587286]=note: print w/replies, xml ) Need Help??


in reply to Re: 15 billion row text file and row deletes - Best Practice?
in thread 15 billion row text file and row deletes - Best Practice?

Sorting 15 billion rows of text will be non trivial in terms of time and/or memory ;). I would drop the deleted SNs as you find them though.

--Brig

  • Comment on Re^2: 15 billion row text file and row deletes - Best Practice?

Replies are listed 'Best First'.
Re^3: 15 billion row text file and row deletes - Best Practice?
by OfficeLinebacker (Chaplain) on Dec 03, 2006 at 17:03 UTC
    Brig,

    You're right. gzipping would be faster than sorting, though, right? Worth it?

    The thought of using DBD::CSV also crossed my mind. That would be using a DB interface to a CSV file. Don't know if that meets the OP's requirement not to use a DB.

    Great topic. ++

    _________________________________________________________________________________

    I like computer programming because it's like Legos for the mind.

Log In?
Username:
Password:

What's my password?
Create A New User
Domain Nodelet?
Node Status?
node history
Node Type: note [id://587286]
help
Chatterbox?
and the web crawler heard nothing...

How do I use this?Last hourOther CB clients
Other Users?
Others studying the Monastery: (2)
As of 2024-04-16 21:26 GMT
Sections?
Information?
Find Nodes?
Leftovers?
    Voting Booth?

    No recent polls found