I have a question about how the experimental sigmas are affected when
one includes resolution shells containing mostly unobserved
reflections. Does this vary with the data reduction software being
used?
One thing I've noticed when scaling data (this with d*trek (Crystal
Clear) since it's the program I use most) is that I/sigma(I) of
reflections can change significantly when one changes the high
resolution cutoff.
If I set the detector so that the edge is about where I stop seeing
reflections and integrate to the corner of the detector, I'll get a
dataset where I/sigma(I) is really compressed - there is a lot of
high resolution data with I/sigma(I) about 1, but for the lowest
resolution shell, the overall I/sigma(I) will be maybe 8-9. If the
data set is cutoff at a lower resolution (where I/sigma(I) in the
shell is about 2) and scaled, I/sigma(I) in the lowest resolution
shell will be maybe 20 or even higher (OK, there is a different
resolution cutoff for this shell, but if I look at individual
reflections, the trend holds). Since the maximum likelihood
refinements use sigmas for weighting this must affect the
refinement. My experience is that interpretation of the maps is
easier when the cut-off datasets are used. (Refinement is via refmac5
or shelx). Also, I'm mostly talking about datasets from well-
diffracting crystals (better than 2 A).
Sue
On Mar 22, 2007, at 2:29 AM, Eleanor Dodson wrote:
> I feel that is rather severe for ML refinement - sometimes for
> instance it helps to use all the data from the images, integrating
> right into the corners, thus getting a very incomplete set for the
> highest resolution shell. But for exptl phasing it does not help
> to have many many weak reflections..
>
> Is there any way of testing this though? Only way I can think of to
> refine against a poorer set with varying protocols, then improve
> crystals/data and see which protocol for the poorer data gave the
> best agreement for the model comparison?
>
> And even that is not decisive - presumably the data would have come
> from different crystals with maybe small diffs between the models..
> Eleanor
>
>
>
> Shane Atwell wrote:
>>
>> Could someone point me to some standards for data quality,
>> especially for publishing structures? I'm wondering in particular
>> about highest shell completeness, multiplicity, sigma and Rmerge.
>>
>> A co-worker pointed me to a '97 article by Kleywegt and Jones:
>>
>> _http://xray.bmc.uu.se/gerard/gmrp/gmrp.html_
>>
>> "To decide at which shell to cut off the resolution, we nowadays
>> tend to use the following criteria for the highest shell:
>> completeness > 80 %, multiplicity > 2, more than 60 % of the
>> reflections with I > 3 sigma(I), and Rmerge < 40 %. In our
>> opinion, it is better to have a good 1.8 Å structure, than a poor
>> 1.637 Å structure."
>>
>> Are these recommendations still valid with maximum likelihood
>> methods? We tend to use more data, especially in terms of the
>> Rmerge and sigma cuttoff.
>>
>> Thanks in advance,
>>
>> *Shane Atwell*
>>
Sue Roberts
Biochemistry & Biopphysics
University of Arizona
[log in to unmask]
|