Hi Clemens

OK so you're saying use only the reflections that are in common between all
datasets and keep the parameterisation the same.  There are clearly two
distinctly different ways in which datasets can differ: either a different
set of indices due to different resolution cut-offs or completeness, and/or
as you say different sets of values of Iobs/Fobs for the same set of
indices.  In practice of course it's very likely to be a mixture of the two
cases.  Clearly your suggestion could be of help in identifying the "best"
dataset in the latter case, but not in the former case, since it's not even
looking at the reflections that are not in common, so there's no way of
knowing whether they would have improved the map had they been used
(without obviously eye-balling the map!).

I just have an uncomfortable feeling about using a data-derived model to
judge the quality of the data: it sounds like a circular argument to me!
IMO there should be two completely distinct steps: first decide on the
"best" data using data-related criteria alone (Rmeas, CChalf, resolution,
whatever), then with this "best" dataset decide on the "best" model using
the model selection metrics (Rfree, free likelihood and their relatives).

Very interesting discussion!

Cheers

-- Ian


On Thu, 4 Nov 2021 at 16:00, Clemens Vonrhein <vonrh...@globalphasing.com>
wrote:

> On Wed, Nov 03, 2021 at 12:54:00PM +0000, Ian Tickle wrote:
> > Suppose you had to compare two datasets differing only in their
> > high-resolution cut-offs.  Now Rwork, Rfree and Rall will inevitably have
> > higher values at the high d* end, which means that if you apply a cut-off
> > at the high d* end all the overall R values will get smaller, so use of
> any
> >  R value as a data-quality metric will tend to result in selection of the
> > lowest-resolution dataset from your candidate set, which may well give a
> lower
> > quality map: not what you want!
>
> Couldn't that be avoided by using the common set of reflection data
> present in both the A and B dataset (from David's example - or even
> across a whole series A-Z)? When at the same time we keep the same
> model parametrisation (i.e. not adding altConfs or waters, keeping
> same TLS definitions etc) it might be useful to compare different data
> processing procedures via the R{free,all,work} during refinement.
>
> As far as I can see, the only difference would be the actual values of
> F and sig(F) (or I and sig(I)) attached to the identical set of Miller
> indices ... right?
>
> Cheers
>
> Clemens
>

########################################################################

To unsubscribe from the CCP4BB list, click the following link:
https://www.jiscmail.ac.uk/cgi-bin/WA-JISC.exe?SUBED1=CCP4BB&A=1

This message was issued to members of www.jiscmail.ac.uk/CCP4BB, a mailing list 
hosted by www.jiscmail.ac.uk, terms & conditions are available at 
https://www.jiscmail.ac.uk/policyandsecurity/

Reply via email to