in reply to Re: selecting columns from a tab-separated-values file
in thread selecting columns from a tab-separated-values file
Replies are listed 'Best First'. | |
---|---|
Re^3: selecting columns from a tab-separated-values file
by Lotus1 (Vicar) on Jan 24, 2013 at 16:47 UTC | |
It turns out that BrowserUK's approach was an order of magnitude quicker than the other approaches presented here. I took the liberty of coding up Sundialsvc4's suggestion of buffering a few hundred thousand lines worth of data before printing them out to a file. I used array refs since there was no reason to use hash refs. I used a Powershell script to time the different approaches.
Here are the results using a 1Gb test file on an idle server with 16 cores and 8Gb RAM: update: (Windows Server 2008 R2)
Here is the sundialsvc4 approach that I put together for the test:
Here is the Kenosis/choroba approach.
| [reply] [d/l] [select] |
by mildside (Friar) on Jan 24, 2013 at 22:59 UTC | |
Great job there Lotus1. I'm curious about the use of splice to clear the array in your code, as below: splice( @$aref );I would probably have used the below: @$aref = ();Is splice faster or better in some other way? | [reply] [d/l] [select] |
by Lotus1 (Vicar) on Jan 25, 2013 at 02:23 UTC | |
Thanks. I started with assigning the empty array but I had a bug somewhere so I stuck the splice in and got it working. I think I forgot the '@' in the first try but put it in with splice. I don't know which is faster but it is only called a handful of times in this approach anyway. | [reply] |
Re^3: selecting columns from a tab-separated-values file
by BrowserUk (Patriarch) on Jan 22, 2013 at 17:15 UTC | |
taking care that I/O doesn’t sneak up on you from the backside in the form of virtual-memory paging. You and your hobby horses. Virtual memory limits don't enter into it. When running, those two programs I posted use 2.7 MB and 3.2 MB respectively when using the 256 times larger read size than standard that I suggest. Even if I increase that 10-fold to 10MB each -- which slows the processing down -- they use a whole 12 MB & 16 MB. The programmer that runs my heating system could handle that. I’m personally not sure that threads would help here I'm not sure that standing on one leg whilst drinking blood from a freshly decapitated chicken would help; so I don;t mention it. And your Approximately 2/3rds of the throughput gains from my posted solution come exactly because the CPU intensive part of the processing -- the spliting and joining of the records -- can run flat out (100% utilisation) on one CPU, whilst the first thread doing the Input is instantly ready to respond to the completion of each read because it isn't having to perform any CPU intensive processing on each record. push it onto an array (of hashrefs). The input is a stream; the output is a stream; What on earth do you need an array of hashrefs for? I suspect that you will be astonished at what just-this does for the program. No. I can pretty much call it without trying it. It will run 3 to 5 times slower than a simple:
That is, instead of taking the OPs 5+ hours to run it will take closer to 24 hours. Why? You'll never know unless you try it. And you won't do that. (And even if you get someone else to do it for you, you won't post the results, because it will show your 'advice' to be nothing more than groundless guessing.) With the rise and rise of 'Social' network sites: 'Computers are making people easier to use everyday'
Examine what is said, not who speaks -- Silence betokens consent -- Love the truth but pardon error.
"Science is about questioning the status quo. Questioning authority".
In the absence of evidence, opinion is indistinguishable from prejudice.
| [reply] [d/l] |
|