the first step is going to be to reduce the 8 million lines to what we're guessing is 400,000 unique sets of characteristics,
How long wil it take you to do that reduction?
Because with the additional efficiencies outlined in 760218 & 760226, and a little threading or forking, you could distribute this over your 8 processors and have the full cross product very quickly.
Of course, that would be a huge amount of data to further process, but maybe you could apply some (cheap) selection criteria prior to output.
Anyway, good luck!
In reply to Re^3: Huge data file and looping best practices
by BrowserUk
in thread Huge data file and looping best practices
by carillonator
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |