Malcolm,

To verify your build, our 'gold standard' is to inspect the
segmentations and surfaces manually.  Of course, this requires skill
both in neuroanatomy, and what to look for in a typical freesurfer
recon.  We do this with our 'buckner40' (40 subject) and 'adni60' (60
subject) datasets prior to release.  We also compare group analysis runs
between versions to look for statistical differences.

For you purposes though, it should be sufficient to run asegstatsdiff
and aparcstatsdiff comparing the results of a single subject run on your
build and the release build.  you should not see hippocampal volume
differences greater than 2%.  you can also send me the results of these
stats diff files.

Also, I added a check to the libg2c make check
(dev/talairach_avi/test_libg2c) to skip it if SKIP_LIBG2C_TEST is set in
the enviro.  unfortunately, make check doesnt allow continuing if a test
fails (or least i dont know of flag to do so).

N.




On Thu, 2012-01-19 at 09:55 -0600, Malcolm Tobias wrote:
> My foray into custom-built binaries looks promising.  Using the distributed 
> (CentOS4) build of 5.1, I'm running the Bert/Ernie benchmark in ~12 hours.  
> If 
> I build this myself (using 4.1.2 / Red Hat 4.1.2-51) I can run in ~10 hours.  
> If I build the same code using the Intel compilers (12.0.0) with pretty 
> generic options I can get this down to ~9 hours, which is the 10% speed-up 
> that Nick suggested.
> 
> Now the hard part: What steps should I take to validate these custom-built 
> binaries before users start using them for production runs?
> 
> It seems like at a minimum I should run and pass all of 'make check'.  These 
> seem to stop after the first error.  I wonder if it would be better to have 
> them run to completion to get an idea of what all passes/fails?
> 
> -For gcc, it fails on the test for libg2c.  I can understand why you need to 
> test for this before releasing to the world, but is there some way for me to 
> disable this particular test so it doesn't stop the rest of the tests from 
> running?
> 
> -For the Intel compilers, the environment doesn't seem to inherit my 
> LD_LIBRARY_PATH so most of the tests fail early on because of missing 
> libraries.  Any thoughts on how to fix this?
> 
> The 'make check' seems to test for basic functionality, not necessarily the 
> consistency of the results (please correct me if I'm wrong).  To test the 
> validity of the processing, it seems the Bert/Ernie sample data is used.  I'm 
> not sure what parts of the results I should check?  It's been suggested that 
> I 
> compare the results of the aseg.stats and aparc.stats files, possibly using 
> asegstatsdiff and aparcstatsdiff.  Does this seem reasonable?  If so, what 
> levels of differences would be considered acceptable?
> 
> To put this all another way, what do you guys normally do to test out a build 
> before releasing it?
> 
> Cheers,
> Malcolm
> 

_______________________________________________
Freesurfer mailing list
Freesurfer@nmr.mgh.harvard.edu
https://mail.nmr.mgh.harvard.edu/mailman/listinfo/freesurfer


The information in this e-mail is intended only for the person to whom it is
addressed. If you believe this e-mail was sent to you in error and the e-mail
contains patient information, please contact the Partners Compliance HelpLine at
http://www.partners.org/complianceline . If the e-mail was sent to you in error
but does not contain patient information, please contact the sender and properly
dispose of the e-mail.

Reply via email to