Print

Print


I'm not sure I entirely agree with ZO's assessment that a B factor is
a measure of uncertainty.  Pedantically, all it really is is an
instruction to the refinement program to "build" some electron density
with a certain width and height at a certain location.  The result is
then compared to the data, parameters are adjusted, etc.  I don't
think the B factor is somehow converted into an "error bar" on the
calculated electron density, is it?

For example, a B-factor of 500 on a carbon atom just means that the
"peak" to build is ~0.02 electron/A^3 tall, and ~3 A wide (full width
at half maximum).  By comparison, a carbon with B=20 is 1.6
electrons/A^3 tall and ~0.7 A wide (FWHM).  One of the "bugs" that
Dale referred to is the fact that most refinement programs do not
"plot" electron density more than 3 A away from each atomic center, so
a substantial fraction of the 6 electrons represented by a carbon with
B=500 will be sharply "cut off", and missing from the FC calculation.
Then again, all 6 electrons will be missing if the atoms are simply
not modeled, or if the occupancy is zero.

The point I am trying to make here is that there is no B factor that
will make an atom "go away", because the way B factors are implemented
is to always conserve the total number of electrons in the atom, but
just spread them out over more space.

Now, a peak height of 0.02 electrons/A^3 may sound like it might as
well be zero, especially when sitting next to a B=20 atom, but what if
all the atoms have high B factors?  For example, if the average
(Wilson) B factor is 80 (like it typically is for a ~4A structure),
then the average peak height of a carbon atom is 0.3 electrons/A^3,
and then 0.02 electrons/A^3 starts to become more significant.  If we
consider a ~11 A structure, then the average atomic B factor will be
around 500.  This "B vs resolution" relationship is something I
derived empirically from the PDB (Holton JSR 2009).  Specifically, the
average B factor for PDB files at a given resolution "d" is: B =
4*d^2+12.  Admittedly, this is "on average", but the trend does make
physical sense: atoms with high B factors don't contribute very much
to high-angle spots.

More formally, the problem with using a high B-factor as a "flag" is
that it is not resolution-general.  Dale has already pointed this out.

Personally, I prefer to think of B factors as a atom-by-atom
"resolution" rather than an "error bar", and this is how I tell
students to interpret them (using the B = 4*d^2+12 formula).  The
problem I have with the "error bar" interpretation is that
heterogeneity and uncertainty are not the same thing.  That is, just
because the atom is "jumping around" does not mean you don't know
where the centroid of the distribution is.  The "u_x" in
B=8*pi^2*<u_x^2> does reflect the standard error of atomic position in
a GIVEN unit cell, but since we are averaging over trillions of cells,
the "error bar" on the AVERAGE atomic position is actually a great
deal smaller than "u".  I think this distinction is important because
what we are building is a model of the AVERAGE electron density, not a
single molecule.

Just my 0.02 electrons

-James Holton
MAD Scientist



On Fri, Apr 1, 2011 at 10:57 AM, Zbyszek Otwinowski
<[log in to unmask]> wrote:
> The meaning of B-factor is the (scaled) sum of all positional
> uncertainties, and not just its one contributor, the Atomic Displacement
> Parameter that describes the relative displacement of an atom in the
> crystal lattice by a Gaussian function.
> That meaning (the sum of all contributions) comes from the procedure that
> calculates the B-factor in all PDB X-ray deposits, and not from an
> arbitrary decision by a committee. All programs that refine B-factors
> calculate an estimate of positional uncertainty, where contributors can be
> both Gaussian and non-Gaussian. For a non-Gaussian contributor, e.g.
> multiple occupancy, the exact numerical contribution is rather a complex
> function, but conceptually it is still an uncertainty estimate. Given the
> resolution of the typical data, we do not have a procedure to decouple
> Gaussian and non-Gaussian contributors, so we have to live with the
> B-factor being defined by the refinement procedure. However, we should
> still improve the estimates of the B-factor, e.g. by changing the
> restraints. In my experience, the Refmac's default restraints on B-factors
> in side chains are too tight and I adjust them. Still, my preference would
> be to have harmonic restraints on U (square root of B) rather than on Bs
> themselves.
> It is not we who cram too many meanings on the B-factor, it is the quite
> fundamental limitation of crystallographic refinement.
>
> Zbyszek Otwinowski
>
>> The fundamental problem remains:  we're cramming too many meanings into
> one number [B factor].  This the PDB could indeed solve, by giving us
> another column.  (He said airily, blithely launching a totally new flame
> war.)
>> phx.
>>
>