in reply to Re^3: Strategy for managing a very large database with Perl (Video)
in thread Strategy for managing a very large database with Perl
BrowserUk> I dummied up two days worth of data files:Could you please share your code to create the dummy data? One, I would see what your data look like, and two, I would learn a few tricks meself.
BrowserUk> With compression, that could be as little as 1.3 TB. BrowserUk> Though you'd have a pay the price for unpacking-- BrowserUk> ~30 seconds per file.RIght. Good to keep in mind, but, not interested in paying the 30 seconds price for optimizing for space. Space is not a constraining factor, for now.
BrowserUk> But the main point of partitioning your dataset BrowserUk> this way is that you reduce the search space to BrowserUk> 1/8th of 1% as soon as you specify the year/day. BrowserUk> And there is no searching of indexes involved in BrowserUk> the rest of the query. Just a simple calculation BrowserUk> and a direct seek.Of course, I have to add the cost of looking up the spatial extent, which I have to do via the Pg database first. But, that can give me the area I want to pluck out of my image, and then work with it.
BrowserUk> Anyway, t'is your data and your employers money :)True. This is research, so trying out different ways is a worthwhile exercise in itself. I don't get paid too much, so it is not a lot of money on my employer's part. ;-). Nevertheless, thanks much. This is fantastic. Always a pleasure to ask and learn different approaches to solving a problem.
|
|---|
| Replies are listed 'Best First'. | |
|---|---|
|
Re^5: Strategy for managing a very large database with Perl (Video)
by BrowserUk (Patriarch) on Jun 18, 2010 at 15:45 UTC |