It seems to be widely known and observed that diffuse background
scattering decreases more rapidly with increasing detector-to-sample
distance than Bragg reflections. For example, Jim Pflugrath, in his
1999 paper (Acta Cryst 1999 D55 1718-1725) says "Since the X-ray
background falls off as the square of the distance, the expectation
is that a larger crystal-to-detector distance is better for reduction
of the x-ray background. ..."
Does anyone know of a more rigorous discussion of why background
scatter fades while Bragg reflections remain collimated with distance?
Richard Gillilan
MacCHESS
|