in reply to advice needed for processing largish data in multiple files

I'd advocate using DB_File (or BerkeleyDB) and tied hashes. That should be faster than DB insert / query processing.

--shmem

_($_=" "x(1<<5)."?\n".q·/)Oo.  G°\        /
                              /\_¯/(q    /
----------------------------  \__(m.====·.(_("always off the crowd"))."·
");sub _{s./.($e="'Itrs `mnsgdq Gdbj O`qkdq")=~y/"-y/#-z/;$e.e && print}
  • Comment on Re: advice needed for processing largish data in multiple files

Replies are listed 'Best First'.
Re^2: advice needed for processing largish data in multiple files
by tilly (Archbishop) on Aug 19, 2006 at 03:53 UTC
    The performance difference between DB_File and DBD::SQLite is minimal. Besides, with this data volume, either will be very fast. The more important consideration is whether the programming problem is one which is naturally expressed with SQL.