I ran such a dictionary over a gigabyte of text on my computer, and it took five minutes to check every word. This works out to about 3.5 days on the whole 1 TB corpus,
It seems unlikely that your slow timings are due to the hash lookup. Far more likely to be down to how you are breaking your data into words?
In reply to Re: Advice for optimizing lookup speed in gigantic hashes
by BrowserUk
in thread Advice for optimizing lookup speed in gigantic hashes
by tobek
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |