"Depends" is about as good of an answer as I could think of, too. (Whenever someone uses a comparative like "better", I usually insist on "better, for what?" if I'm going to spend time answering, and start thinking "is this an X-Y question?", etc.) I would say that I don't usually use selectall_hashref like that. Often I'd go with selectall_arrayref($sql, {Slice => {}}, @bind) so you iterate over the hrefs directly to build the aggregations (hashes). And end up, like you say, splitting things up and doing subqueries. Depends..
| [reply] [Watch: Dir/Any] [d/l] [select] |
Yep.
I will traverse all the data
The data lives on another server
Pulling it all at once is just faster*
I acutlly have both selectall_hashref and selectall_arrayref($sql, {Slice => {}}) in my code, each where it is best (I believe)
* I should probably benchmark that
| [reply] [Watch: Dir/Any] [d/l] [select] |
Have you given thought that maybe you should just do the work on the db server via stored procedure or some such, and only return back the data needed to render user output (regardless of format)? Since you've claimed both of these keys to be unique, you presumably already have enforced that in the db and that should mean they're indexed anyway, and then you don't really need to worry about how many lookup keys you have.
And, out of left field, comes Tanktalus...
| [reply] [Watch: Dir/Any] |
So "best" means fastest?
We don't know how big your table is and if memory is an issue.
In programming you can almost always trade memory with time!
... like pulling n big chunks of the table in sliding windows.
For instance I suppose (see your footnote) two SQL queries for ID and SN are faster but your map solution occupies less memory ( the second level hash refs are reused scalars)
So .... It really depends...
>
* I should probably benchmark that
That's a bingo! ;)
| [reply] [Watch: Dir/Any] |