Beefy Boxes and Bandwidth Generously Provided by pair Networks
Keep It Simple, Stupid

Re: Parse data into large number of output files.

by davido (Cardinal)
on Sep 29, 2004 at 01:42 UTC ( #394816=note: print w/replies, xml ) Need Help??

in reply to Parse data into large number of output files.

Perhaps you could do this in a couple of stages. In stage one, read all of the input files, and build up a database with a column for "trap sender" and a column for the logfile entry, each row in the DB representing an entry from one of the input logfiles. It takes only one database connection to build up the database.

Then in the second pass, prepare a query such as "SELECT logentry FROM sometable WHERE trap_sender=?". Then one by one open a file for a particular user, execute the query with that user's name, and spill to that user's logfile the entries returned by fetchrow.

That way, you never have to hold all 100mb in memory at once, you never have to hold hundreds of open filehandles, you don't have to suffer the poor performance of opening and closing filehandles hundreds of times over and over again, and you get some practice with DBI. DBD::SQLite would be a great lightweight database on which to build the implementation of this strategy.


  • Comment on Re: Parse data into large number of output files.

Log In?

What's my password?
Create A New User
Node Status?
node history
Node Type: note [id://394816]
and the web crawler heard nothing...

How do I use this? | Other CB clients
Other Users?
Others chanting in the Monastery: (7)
As of 2020-11-24 21:54 GMT
Find Nodes?
    Voting Booth?

    No recent polls found