Re^3: Reduce CPU utilization time in reading file using perlby BrowserUk (Patriarch)
|on Sep 28, 2013 at 13:31 UTC
I have tried your option as well, sitll the CPU usage is 100%.
That is because you are using more memory for the hash than you have installed, thus, some parts of the memory holding the hash are being swapped or paged to disk as the file is being read. The nature of the way hashes are stored means that pages of memory are constantly being written to disk and then re-read, over and over; and that is what is driving up your cpu usage.
Is it possible to load only one line into memory from the huge file and store it into hash or without opening directly possible to store it into hash.
That is what my code does. Read's one line installs into the hash then reads the next. It is the size of the hash that is the problem, not the line-by-line processing of the file.
I am worrying that it may not be very less possible. Still thought of getting your suggestion.
There are various ways of providing access to huge amount of data without requiring that it all be held in memory concurrently. Which of those methods/mechanisms is appropriate for your purpose depends entirely upon what you need to do with that data.
So, before advising further, you need to answer the question: Why are you attempting to load all the data into a hash?
With the rise and rise of 'Social' network sites: 'Computers are making people easier to use everyday'
Examine what is said, not who speaks -- Silence betokens consent -- Love the truth but pardon error.
"Science is about questioning the status quo. Questioning authority".
In the absence of evidence, opinion is indistinguishable from prejudice.