Dear all,
I'm a bit confused from the output of the CORRECT step in XDS. In one of the first tables I can read the mean I/sigma for each resolution shell, but these values are much different from the I/sigma reported in the table at the end of the output files, titled "completeness and quality of data set" for the full data range with signal/noise > -3.0. For example, from the first table I have I/sigma = 2 at 3.6 A, while from the second table I have I/sigma = 2 at 2.8 A! What is exactly the difference between the two values? And which one is reliable to decide the resolution cutoff?


Thank you in advance,

Michele Lunelli
MPI for Infection Biology
Berlin - Germany








----

Reply via email to