You can do this with Scala or Aimless. Scale everything first, write out the scaled unmerged file, then read it again "onlymerge" and a batch selection (Aimless also gives a cumulative completeness) Phil
On 12 Jan 2012, at 14:00, Ingo P. Korndoerfer wrote: > hello, > > my poor dementia ridden brain has gone on screensaver ... > > i need to calculate the completeness and redundancy of reflections in > batches or ranges of batches in a > multi-record .mtz file. > > sftools can do this, but the numbers are pretty much meaningless, i.e., > my feeling is, if i measure > 10% of reflections 10 times, it will give me 100% completeness. > > second option is, to select the batches i want, purge the rest, force > sftools to "merge average" > and then ask for the completeness. this works. but it requires > re-reading of the complete > dataset for every batch or segment of batches i am interested in, which > is too slow, > > any ideas greatly appreciated :-) > > 1000 thanks already > > ingo