Dear Crystallographers,

One can see from many posts on this listserve that in any given x-ray diffraction experiment, there are more data than merely the diffraction spots. Given that we now have vastly increased computational power and data storage capability, does it make sense to think about changing the paradigm for model refinements? Do we need to "reduce" data anymore? One could imagine applying various functions to model the intensity observed at every single pixel on the detector. This might be unneccesary in many cases, but in some cases, in which there is a lot of diffuse scattering or other phenomena, perhaps modelling all of the pixels would really be more true to the underlying phenomena? Further, it might be that the gap in R values between high- and low-resolution structures would be narrowed significantly, because we would be able to model the data, i.e., reproduce the images from the models, equally well for all cases. More information about the nature of the underlying macromolecules might really be gleaned this way. Has this been discussed yet?

Regards,

Jacob Keller

*******************************************
Jacob Pearson Keller
Northwestern University
Medical Scientist Training Program
Dallos Laboratory
F. Searle 1-240
2240 Campus Drive
Evanston IL 60208
lab: 847.491.2438
cel: 773.608.9185
email: j-kell...@northwestern.edu
*******************************************

Reply via email to