in reply to Re: Re: Re: Memory allocation strategies
in thread Memory allocation strategies

Hmm, thanks for the reply BrowserUk. I had thought of something like this, but unfortunately I need the full array in contiguous memory at all times. The problem is that the base algorithm needs to be optimized for lookups, and as such the lookup is done in C and is a very minimal function (all precondition/postcondition checks are handled in the perl layer on STORE so the FETCH can be as minimal as possible) I will likely have to call the lookup several million times more than the store. So if I use a discontinuous representation I have to put a bunch of logic in the lookup to handle those cases, or repack the whole thing if a STORE has occured since the last FETCH, and it is precisely the last case that is the reason for me using the implementation I am. Originally I just used a normal array to represent the object from the perl POV and a packed representation (packed on demand when a FETCH occured after a STORE) for the C. However this meant that I ended up storing the data twice, and the array storage was very wasteful of ram.

Anyway, your idea made me think of an alternate strategy for allocation. The likelyhood that the base algorithm needs to extend the array is inversely proportional to the number of items in the data structure for typical data sets for what the algorithm is intended for. So I was thinking of setting up an array of allocation sizes that goes from large, say 64k down to relatively small (4096 sounds good based on the reasons you point out, thanks). This way the initial allocation will cover small data sets nicely without wasting too much ram, and later allocation will be more conservative and still end up "fitting" quite nicely. I havent done this yet, but I plan to have a look at its behaviour and see how it goes.

Cheers and thanks again for the reply. Some good stuff there.


---
demerphq

<Elian> And I do take a kind of perverse pleasure in having an OO assembly language...

Replies are listed 'Best First'.
Re: Re: Re: Re: Re: Memory allocation strategies
by BrowserUk (Patriarch) on Sep 02, 2003 at 11:37 UTC
    If your doing this from C, then you might consider taking both Perl and the C-runtime out of the equation completely and going directly to the OS for your memory.

    The OS has an API with which you can reserve a large amount of (virtual) memory without actually allocating it. You then make a second call to actually commit chunks of the reserved allocation as you need it. In this way, you can reserve a contiguous address space large enough to accomodate your biggest likely requirements and then grow the actual allocation into that reserved contiguous address space as you need to. This removes the need for any reallocation and the copying and (temporary) duplication that that entails and bypasses several intermediate levels of freespace chain chasing, coalesing and shuffling to boot.

    As your search code is written in C, there is no great problem with utilising the address space provided by the OS for this purpose, you don't need for the overall space to be managed by Perl. You can simple copy your individual array elements to a Perl scalar for returning to the perl code once you have located it.

    Needless to say I haven't done this from perl, but the C part is relatively trivial. I located this brief description and sample code on MSDN that demonstrates using this technique if the idea is of any interest.


    Examine what is said, not who speaks.
    "Efficiency is intelligent laziness." -David Dunham
    "When I'm working on a problem, I never think about beauty. I think only how to solve the problem. But when I have finished, if the solution is not beautiful, I know it is wrong." -Richard Buckminster Fuller
    If I understand your problem, I can solve it! Of course, the same can be said for you.