On Sat, 11 Nov 2006 23:01:30 -0800
"Will Fould" <[EMAIL PROTECTED]> wrote:

> I've found conflicting information about this.
> 
> modperl: Is there a performance impact to breaking up a single 350k
> file (monolithic package) into multiple files (still the same package
> name)?
> 
> I guess I'm asking: Are larger files inherently slower, even when
> pre-compiled?  (Obviously, I'm not a computer science student)
> 
> Here's the situation:
> 
> Over time, what was a 60k package has grown to 350k, encompassing
> many subs that do more or less,  3 mutually exclusive operations.   I
> want to remove 70% of the subs and put them into their own files for
> maintenace/organization reasons, but is there a performance benefit
> or hit by doing this?  The main file will basically direct,
> requireing the additional files only when necessary. But under mod
> perl, I believe the require will only happen once (per process)
> anyway.

    You won't see any real performance improvement, but it certainly
    would be an improvement to your code.  

    No offense, but I've never seen a 350k module that didn't need
    to be broken into SEVERAL other modules for
    readability/maintainability reasons.  

    I'm over generalizing here and not taking into account
    bloated/memory hogging modules, but files/packages/modules
    are cheap to use and create and should be broken up for reasons
    other than performance.  

    I just looked over one of the largest mod_perl apps I help 
    maintain, there are 67 modules and only one of them is over 
    20k in size. 

 ---------------------------------
   Frank Wiles <[EMAIL PROTECTED]>
   http://www.wiles.org
 ---------------------------------

Reply via email to