Hi all,

sorry for the double post, but I noticed an error in the chosen title.

I believe the website warns for high demands of memory, claiming it may
take up to 10G for this processing step. However, when I run this on
cluster the memory demands go up to 30G. This is still for one subject
using one scan. Is this to be expected, has something changed in the
processing pipeline?

Eventually I want to run the longitudinal version, should I expect an
increase in memory demands?

Running the parahippocampal subfields on our cluster using:
recon-all -s $subject -hippocampal-subfields-T1

Version:
INFO: FreeSurfer build stamps do not match
Subject Stamp: freesurfer-Linux-centos6_x86_64-stable-v6-beta-20161028-
4cfbf2a
Current Stamp: freesurfer-Linux-centos6_x86_64-stable-pub-v6.0.0-2beb96c

Kind regards,

Ferdi
_______________________________________________
Freesurfer mailing list
Freesurfer@nmr.mgh.harvard.edu
https://mail.nmr.mgh.harvard.edu/mailman/listinfo/freesurfer


The information in this e-mail is intended only for the person to whom it is
addressed. If you believe this e-mail was sent to you in error and the e-mail
contains patient information, please contact the Partners Compliance HelpLine at
http://www.partners.org/complianceline . If the e-mail was sent to you in error
but does not contain patient information, please contact the sender and properly
dispose of the e-mail.

Reply via email to