IMHO, this is not a computationally difficult task. It really boils down to first loading the reference data into a hash and, then, read a single input file and lookup into the hash. Quite simple. And since there is apparently only one data input file, I doubt that using threads will bring any performance benefit.
In reply to Re^2: How to make a hash to evaluate columns between large datasets
by Laurent_R
in thread How to make a hash to evaluate columns between large datasets
by rambosauce
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |