JiscMail Logo
Email discussion lists for the UK Education and Research communities

Help for CCPNMR Archives


CCPNMR Archives

CCPNMR Archives


CCPNMR@JISCMAIL.AC.UK


View:

Message:

[

First

|

Previous

|

Next

|

Last

]

By Topic:

[

First

|

Previous

|

Next

|

Last

]

By Author:

[

First

|

Previous

|

Next

|

Last

]

Font:

Monospaced Font

LISTSERV Archives

LISTSERV Archives

CCPNMR Home

CCPNMR Home

CCPNMR  January 2008

CCPNMR January 2008

Options

Subscribe or Unsubscribe

Subscribe or Unsubscribe

Log In

Log In

Get Password

Get Password

Subject:

Re: isotope-assignment mismatch: Resetting isotope

From:

Tim Stevens <[log in to unmask]>

Reply-To:

CcpNmr software mailing list <[log in to unmask]>

Date:

Mon, 7 Jan 2008 15:24:04 +0000

Content-Type:

MULTIPART/MIXED

Parts/Attachments:

Parts/Attachments

TEXT/PLAIN (55 lines) , importFromCcpn.py (1 lines) , Protocol.py (1 lines)


Attached are two more files for ARIA's src/py/ directory. The
importFromCcpn.py fixes the transfer of ambiguous prochiral constraints to
ARIA. (Without this HBa may come back as HBa/HBb in the final restraints).
Protocol.py is needed if you wan to calibrate CCPN constraint lists.


> I have just finished my new Aria run with exportToCCPN.py updated. When I
> loaded the CCPNMR project I still have the error messages.

I have been running a great deal ARIA2 recently and the dim/isotope
mismatch issue has never reappeared for me. If you are still getting
errors then the solution may be remove all/any old ARIA generated
peaklists; even if the new export is OK, the remanents of the old export
may remain.


> And while I looked inside my unambig.tbl I have dicovered this line:
>
> assign (segid " A" and resid 96 and name HN) (segid " A" and resid
> 94 and name HB) 4.822 0.634 0.634 weight 1.000 ! spec=3daroCnoesy,
> no=169, id=85, vol=1.628858e+06
>
> Thus in my 3daroCnoesy spectrum Aria2.2 makes assignment of CH peaks to an
> amide group.

This looks like the sort of error that the old code would cause (mismatch
between the restraint and the original CCPN peak, and hence the spectrum)
So, I think there may be some of the old erroneous data lying about
somewhere.


> So I wonder how people work with CCPNMR/Aria to get around this bug. Can
> we believe unambig.tbl and ambig.tbl, just delete the wrong lines and
> load the tables into CCPNMR with FormatConverter as Tim suggested? If we
> can, what are the next steps people do in CCPNMR for NOESY spectra
> assignment?

If possible, I would go back to state and ARIA run before the erroneous
export. Everything really is working fine for me here.

T.

-------------------------------------------------------------------------------
 Dr Tim Stevens Email: [log in to unmask]
 Department of Biochemistry [log in to unmask]
 University of Cambridge Phone: +44 1223 766018 (office)
 80 Tennis Court Road +44 7816 338275 (mobile)
 Old Addenbrooke's Site +44 1223 364613 (home)
 Cambridge CB2 1GA WWWeb: http://www.bio.cam.ac.uk/~tjs23
 United Kingdom http://www.pantonia.co.uk
-------------------------------------------------------------------------------
------ +NH3CH(CH(CH3)OH)C(O)NHCH(CH(CH3)CH2CH3)C(O)NHCH(CH2CH2SCH3)CO2- -------
-------------------------------------------------------------------------------


from Molecule import Molecule
from Chain import TYPE_PROTEIN, TYPE_DNA, TYPE_RNA, TYPE_NONPOLYMER, Chain, ChainSettings
from Residue import Residue
from Atom import Atom, AtomSettings
from ccpn2top import setupEquivalentGroup


from Assignment import Assignment, ASSIGNMENT_TYPE_MANUAL
from NOESYSpectrum import NOESYSpectrum
from CrossPeak import CrossPeak
from Datum import ChemicalShift, Datum
from ChemicalShiftList import ChemicalShiftList

from ShiftAssignment import AVERAGING_METHOD_FAST, AVERAGING_METHOD_SLOW, AVERAGING_METHOD_NONE,\
ASSIGNMENT_METHOD_STEREO_SPECIFIC, ASSIGNMENT_METHOD_EQUIVALENT, \
ASSIGNMENT_METHOD_FLOATING, SpinSystem, ShiftAssignment

from tools import string_to_segid
from ccpnmr.format.general.Constants import ccpNmr_kw, isAriaInput_kw

from memops.api import Implementation
from memops.universal.Constants import True

from Settings import Settings
from xmlutils import XMLBasePickler

chainTypeMapping = {'protein' : TYPE_PROTEIN,
'DNA' : TYPE_DNA,
'RNA' : TYPE_RNA,
'nonpolymer': TYPE_NONPOLYMER}

def getKeysFromString(word, delimiter='|'):
"""Descrn: Get a list of CCPN data model keys given an object identifier string.
Inputs: String
Output: List of Keys (Words or Ints)
"""

items = word.split(delimiter)

keys = []
for item in items:

try:
key = int(item)
except:
key = item

## if item[0] in "0123456789":
## key = int(item)

## else:
## key = item

keys.append(key)

return keys


def getObjectKeys(object):
"""Descrn: Get a list of CCPN data model keys for a CCPN data model object.
Inputs: CCPN data model object
Output: List of Keys (Words or Ints or a further list of keys)
"""

return object.getExpandedKey()


def getObjectKeyString(object, delimiter='|'):
"""Descrn: Make an object identifier string for a CCPN data model object
Inputs: CCPN data model object
Output: String
"""

keys = object.getExpandedKey()

for i in range(len(keys)):
key = keys[i]

keyType = type(key)
if keyType is type([]):
keys[i] = delimiter.join([str(k) for k in key])
elif keyType is not type(''):
keys[i] = str(key)

return delimiter.join(keys)

# BARDIAUX
def getObjectsKeyString(o):

s = []
for a in o:
s.append(getKeysFromString(getObjectKeyString(a)))

base = s[0]

for i,j in zip(*s):
if i <> j:
base += j

return "|".join(base)

def getShiftLists(project):
"""Descrn: Get a list of shift lists from a CCPN project
Inputs: Implementation.Project
Output: List of ccp.nmr.Nmr.ShiftLists
"""

nmrProject = project.currentNmrProject

shiftLists = []
for measurementList in nmrProject.sortedMeasurementLists():
if measurementList.className == 'ShiftList':
shiftLists.append(measurementList)

return shiftLists

def getNoesyPeakLists(project, molSystem=None):
"""Descrn: Get the NOE peak lists from a CCPN project. Can filter if appropriate to a given
molecular system if passed in.
Inputs: Implementation.Project, ccp.molecule.MolSystem.MolSystem
Output: List of ccp.nmr.Nmr.Peaks
"""

peakLists = []
for experiment in project.currentNmrProject.experiments:
if experiment.molSystems:
if molSystem and (molSystem not in experiment.molSystems):
continue

if experiment.refExperiment:
if experiment.refExperiment.nmrExpPrototype.name.find('NOESY') < 0:
continue
else:
transfer = experiment.findFirstExpTransfer(transferType='NOESY')
if not transfer:
continue

for spectrum in experiment.dataSources:
if (spectrum.dataType == 'processed') and (spectrum.numDim > 1):

isotopes = []
for dataDim in spectrum.dataDims:
for expDimRef in dataDim.expDim.expDimRefs:
if expDimRef.measurementType in ('shift','Shift'):
isotope = ','.join(expDimRef.isotopeCodes)
isotopes.append(isotope)
break

if isotopes.count('1H') > 1:
for peakList in spectrum.peakLists:
if peakList.findFirstPeak():
peakLists.append(peakList)


return peakLists


def getCcpnPeakList(ccpnProject, peakListKey):
"""Descrn: Fetch a CCPN peak list from a CCPN project using a list of object keys
Inputs: Implementation.Project, List of keys (Words or Ints)
Output: ccp.nmr.Nmr.PeakList
"""

nmrProjectName, exptSerial, specSerial, plSerial = peakListKey

nmrProject = ccpnProject.findFirstNmrProject(name=nmrProjectName)
if not nmrProject:
ValueError('No NMR project with name %s' % nmrProjectName)

experiment = nmrProject.findFirstExperiment(serial=exptSerial)
if not experiment:
ValueError('Could not find experiment %d in CCPN project.' % (exptSerial))
return

spectrum = experiment.findFirstDataSource(serial=specSerial)
if not spectrum:
ValueError('Could not find spectrum %d in experiment %d.' % (specSerial, exptSerial))
return

peakList = spectrum.findFirstPeakList(serial=plSerial)
if not peakList:
ValueError('Could not find peak list serial %d in spectrum %d:%d' % (plSerial, exptSerial, specSerial))

return peakList


def getCcpnShiftList(ccpnProject, shiftListKey):
"""Descrn: Fetch a CCPN shift list from a CCPN project using a list of object keys
Inputs: Implementation.Project, List of keys (Words or Ints)
Output: ccp.nmr.Nmr.ShiftList
"""

nmrProjectName, shiftListSerial = shiftListKey

nmrProject = ccpnProject.findFirstNmrProject(name=nmrProjectName)
if not nmrProject:
ValueError('No NMR project with name %s' % nmrProjectName)

shiftList = nmrProject.findFirstMeasurementList(serial=shiftListSerial)
if not shiftList:
ValueError('Could not find shift list %d in CCPN project' % (int(shiftListSerial)))

return shiftList


def getCcpnExperimentShiftList(ccpnExperiment):
"""Descrn: Fetch the CCPN shift list that corresponds to a CCPN NMR experiment
Inputs: ccp.nmr.Nmr.Experiment
Output: ccp.nmr.Nmr.ShiftList
"""

shiftList = ccpnExperiment.shiftList

if not shiftList:
for spectrum in ccpnExperiment.dataSources:
for peakList in spectrum.peakLists:
for peak in peakList.peaks:
for peakDim in peak.peakDims:
for contrib in peakDim.peakDimContribs:
shift = contrib.resonance.findFirstShift()
if shift:
return shift.parentList

return shiftList


def getCcpnChain(ccpProject, chainKey):
"""Descrn: Fetch a CCPN chain from a CCPN project using a list of object keys
Inputs: Implementation.Project, List of keys (Words or Ints)
Output: ccp.molecule.MolSystem.Chan
"""

molSystemCode, chainCode = chainKey

molSystem = ccpProject.findFirstMolSystem(code=molSystemCode)
if not molSystem:
ValueError('No molecular system with code "%s" in CCPN project' % molSystemCode)

chain = molSystem.findFirstChain(code=chainCode)
if not chain:
ValueError('No chain found with code "%s" in molecular system "%s"' % (chainCode, molSystemCode))

return chain

# BARDIAUX
def getCcpnChains(ccpProject, chainKey):
"""Descrn: Fetch a CCPN chain from a CCPN project using a list of object keys
Inputs: Implementation.Project, List of keys (Words or Ints)
Output: list of ccp.molecule.MolSystem.Chains
"""

molSystemCode, chainCodes = chainKey[0], chainKey[1:]

molSystem = ccpProject.findFirstMolSystem(code=molSystemCode)
if not molSystem:
ValueError('No molecular system with code "%s" in CCPN project' % molSystemCode)

chains = []
for chainCode in chainCodes:
chain = molSystem.findFirstChain(code=chainCode)
if not chain:
ValueError('No chain found with code "%s" in molecular system "%s"' % (chainCode, molSystemCode))

chains.append(chain)

return chains

def getCcpnPeakAndShiftLists(ccpProject, molSystem, peakShiftListKeys):
"""Descrn: Fetch a list of peak list and corresponding shift list pairs for a CCPN project
given pairs of corresponding object key lists
Inputs: Implementation.Project, List of 2-List of keys (Words or Ints)
Output: List of (ccp.nmr.Nmr.PeakList, ccp.nmr.Nmr.ShiftList)
"""

peakAndShiftLists = []

for peakListKey, shiftListKey in peakShiftListKeys:
shiftlist = None
peakList = getCcpnPeakList(ccpProject, peakListKey)

if peakList:
experiment = peakList.dataSource.experiment
if (experiment.molSystems) and (molSystem not in experiment.molSystems):
data = (molSystem.code, experiment.serial, experiment.name)
print 'WARNING: Selected molecular system "%s" does not match experiment %d (%s) molecular systems' % data

shiftList = getCcpnExperimentShiftList(experiment)

shiftList0 = getCcpnShiftList(ccpProject, shiftListKey)

if shiftList0:
shiftList = shiftList0
if shiftList and (shiftList0 is not shiftList):
print 'WARNING: Selected CCPN shift list does not match experiment %d (%s) shift list' % (experiment.serial, experiment.name)

if peakList and shiftList:
peakAndShiftLists.append([peakList, shiftList])

if not peakAndShiftLists:
print 'ERROR: No selectable CCPN peak lists.'
return []

return peakAndShiftLists

def getCcpnConstraintLists(ccpnProject, restraintsNames):
"""Descrn: Fetch a CCPN constraint list from a CCPN project using list
of restraint type and corresponding object keys
Inputs: Implementation.Project, ARIA restraint names (what object is this?)
Output: List if ccp.nmr.NmrConstraint.AbstractConstraintLists
"""

constraintLists = {}
for restraintType, constraintListKeys in restraintsNames.items():
constraintLists[restraintType] = []

for constraintListKey in constraintListKeys:
keys = getKeysFromString(constraintListKey)
constraintList = getCcpnConstraintList(ccpnProject, keys)

if constraintList:
constraintLists[restraintType].append(constraintList)


return constraintLists


def getCcpnConstraintList(ccpnProject, keys):
"""Descrn: Fetch a CCPN constraint list from a CCPN project using list of object keys
Inputs: Implementation.Project, List of keys (Words or Ints)
Output: ccp.nmr.NmrConstraint.AbstractConstraintList
"""

constraintList = None
nmrProjectKey, storeSerial, serial = keys

nmrProject = ccpnProject.findFirstNmrProject(name=nmrProjectKey)
if not nmrProject:
raise ValueError('No NMR project with name %s' % nmrProjectKey)

constraintStore = nmrProject.findFirstNmrConstraintStore(serial=storeSerial)
if not constraintStore:
raise ValueError('No NMR constraint store with serial "%d" in CCPN project' % storeSerial)
else:
constraintList = constraintStore.findFirstConstraintList(serial=serial)

if constraintList is None:
raise ValueError('No constraint list with serial "%d" in store "%d"' % (serial,storeSerial))

return constraintList


def makeAriaMolecule(ccpMolSystem, ccpChain=None):
"""Descrn: Make an ARIA Molecule given a CCPN MolSystem object
Inputs: ccp.molecule.MolSystem.MolSystem
Output: ARIA Molecule
"""

aria_molecule = Molecule(name=ccpMolSystem.code)

for chain in ccpMolSystem.chains:

if not ccpChain or chain in ccpChain:
aria_chain = makeAriaChain(chain)
aria_molecule.add_chain(aria_chain)

return aria_molecule


def makeAriaChain(ccpChain):
"""Descrn: Make an ARIA Chain given a CCPN Chain object
Inputs: ccp.molecule.MolSystem.Chain
Output: ARIA Chain
"""


# Does below work for DNA/RNA?

aria_settings = ChainSettings()
aria_settings['type'] = chainTypeMapping[ccpChain.molecule.molType]

aria_chain = Chain(settings=aria_settings, segid=string_to_segid(ccpChain.code))

for residue in ccpChain.residues:

aria_residue = makeAriaResidue(residue)
aria_chain.addResidue(aria_residue)

return aria_chain


def makeAriaResidue(ccpResidue):
"""Descrn: Make an ARIA Residue given a CCPN Residue object
Inputs: ccp.molecule.MolSystem.Residue
Output: ARIA Residue
"""


## TODO: have to define chemComp ARIA sysName for residues
## that match, throw error if not recognized (have to define
## own topology file for CNS for now)

try:
aria_residue = Residue(number=ccpResidue.seqCode, residue_type=ccpResidue.ccpCode)

except:
message = 'Residue %s not recognized in ARIA - define topology!'
raise Exception(message % ccpResidue.ccpCode)

ariaAtomDict = {}

idNum = 0
for atom in ccpResidue.atoms:

aria_atom = makeAriaAtom(atom, idNum)
aria_residue.addAtom(aria_atom)
ariaAtomDict[atom.name] = aria_atom
idNum += 1

setupEquivalentGroup(ccpResidue.chemCompVar, aria_residue, ariaAtomDict)

return aria_residue


def makeAriaAtom(ccpAtom, idNum, heteroElements=('N', 'C')):
"""Descrn: Make an ARIA Atom given a CCPN Atom object
Also sets up the hetero atom name.
Inputs: ccp.molecule.MolSystem.Atom
Output: ARIA Atom
"""

chemAtom = ccpAtom.chemAtom
elementSymbol = chemAtom.elementSymbol

heteroName = None
if elementSymbol == 'H':
for bound in chemAtom.findFirstChemBond().chemAtoms:
if bound.elementSymbol in heteroElements:
heteroName = bound.name
break

aria_settings = AtomSettings()
aria_settings['type'] = elementSymbol
aria_settings['hetero_atom_name' ] = heteroName

aria_atom = Atom(settings=aria_settings, name=ccpAtom.name, id=idNum)
aria_atom._setSegid(string_to_segid(ccpAtom.residue.chain.code))

return aria_atom


def makeAriaChemicalShift(ccpShift):
"""Descrn: Make an ARIA ChemicalShift given a CCPN Shift object
Inputs: ccp.nmr.Nmr.Shift
Output: ARIA ChemicalShift
"""

if ccpShift is None:
val = None
err = None
else:
val = ccpShift.value
err = ccpShift.error

return ChemicalShift(val, err)


def makeAriaShiftList(ccpShiftList, ccpMolSystem, ariaMolecule):
"""Descrn: Make a populated ARIA ChemicalShiftList for a given molecule given a CCPN ShiftList object
Inputs: ccp.nmr.Nmr.Shift, ccp.molecule.MolSystem.MolSystem, ARIA Molecule
Output: ARIA ChemicalShiftList
"""

aria_shiftList = ChemicalShiftList()


ss = []

shiftLookup = {}
for shift in ccpShiftList.measurements:
shiftLookup[shift.resonance] = shift

for chain in ccpMolSystem.chains:
for residue in chain.residues:
atomSetDict = {}

for atom in residue.atoms:
atomSet = atom.atomSet
if atomSet and atomSet.resonanceSets:
atomSetDict[atomSet] = True

for atomSet in atomSetDict.keys():
if atomSetDict[atomSet] is False:
continue

resonanceSets = list(atomSet.resonanceSets)

if len(resonanceSets) > 1:
print 'MESSAGE: CCPN atom set %d%s %s has multiple resonance assignments' % (residue.seqCode, residue.ccpCode, atomSet.name)
nonstereo = None
stereo = None
for resonanceSet0 in resonanceSets:
if len(resonanceSet0.atomSets) > 1:
nonstereo = resonanceSet0
else:
stereo = resonanceSet

if stereo:
if nonstereo:
atomSets2 = list(nonstereo.atomSets)
atomSets2.remove(atomSet)
resonanceSet = nonstereo
for resonanceSet0 in atomSets2[0].resonanceSets: # resonance set of other atomSet in pair
if len(resonanceSet0.atomSets) == 1: # if other atom set stereospecifically assigned
resonanceSet = stereo
else:
resonanceSet = stereo
else:
resonanceSet = nonstereo
else:
resonanceSet = resonanceSets[0]

nA = len(resonanceSet.atomSets)
nR = len(resonanceSet.resonances)

if (nA==1) and (nR==1):
ariaAtoms = getAriaAtomsFromResonance(resonanceSet.findFirstResonance(), ariaMolecule)[0]

if len(atomSet.atoms) == 1:
method = AVERAGING_METHOD_NONE
assignment_type = ASSIGNMENT_METHOD_STEREO_SPECIFIC

else:
method = AVERAGING_METHOD_FAST
assignment_type = ASSIGNMENT_METHOD_EQUIVALENT

shift = shiftLookup.get(resonanceSet.findFirstResonance())

if shift:
aria_shift = makeAriaChemicalShift(shift)

aria_spinsystem = SpinSystem(method)
aria_spinsystem.setAtoms(tuple(ariaAtoms))
aria_spinsystem.setChemicalShifts((aria_shift,))

if (aria_spinsystem,) in ss:
continue
ss.append((aria_spinsystem,))

aria_shiftassignment = ShiftAssignment(assignment_type)
aria_shiftassignment.setSpinSystems((aria_spinsystem,))
aria_shiftList.addShiftAssignment(aria_shiftassignment)

atomSetDict[atomSet] = False


elif (nA==2) and (nR<3):

ariaShifts = []
for resonance in resonanceSet.resonances:
shift = shiftLookup.get(resonance)
if shift:
ariaShifts.append(makeAriaChemicalShift(shift))

while len(ariaShifts) < nA:
ariaShifts.append(ChemicalShift(None))

aria_spinsystems = []

for ariaAtoms in getAriaAtomsFromResonance(resonanceSet.findFirstResonance(), ariaMolecule):
if len(ariaAtoms) == 1:
method = AVERAGING_METHOD_NONE

else:
method = AVERAGING_METHOD_FAST

aria_spinsystem = SpinSystem(method)
aria_spinsystem.setAtoms(tuple(ariaAtoms))
aria_spinsystem.setChemicalShifts(tuple(ariaShifts))
aria_spinsystems.append(aria_spinsystem)

atomSetDict[atomSet] = False

if aria_spinsystems in ss:
continue
ss.append(aria_spinsystems)

aria_shiftassignment = ShiftAssignment(ASSIGNMENT_METHOD_FLOATING)
aria_shiftassignment.setSpinSystems(tuple(aria_spinsystems))
aria_shiftList.addShiftAssignment(aria_shiftassignment)


return aria_shiftList


def makeAriaSpectrum(peakList, ariaMolecule):
"""Descrn: Make an ARIA NOESY Spectrum (and assign it)
given a CCPN peak list and ARIA Molecule
Inputs: ccp.nmr.Nmr.PeakList, ARIA Molecule
Output: ARIA NOESYSpectrum
"""

ariaDimNames = ('Proton1','Proton2','Hetero1','Hetero2')

spectrum = peakList.dataSource
experiment = spectrum.experiment
shiftList = experiment.shiftList or experiment.nmrProject.findFirstMeasurementList(className='ShiftList')

transfer = experiment.findFirstExpTransfer(transferType='NOESY')

if not transfer:
raise Exception('Not a NOESY.')

light_dim1 = None
light_dim2 = None
heavy_dim1 = None
heavy_dim2 = None

expDimRefDict = {}

for expDimRef in transfer.sortedExpDimRefs():
if expDimRef.isotopeCodes != ('1H',):
raise Exception('Not an H-H NOESY')

onebondTransfer = expDimRef.findFirstExpTransfer(transferType='onebond')

if onebondTransfer:
expDimRefs = list(onebondTransfer.expDimRefs)
expDimRefs.remove(expDimRef)
expDimRefX = expDimRefs[0]

if light_dim1 is None:
heavy_dim1 = findCcpnDataDim(spectrum, expDimRefX.expDim)
light_dim1 = findCcpnDataDim(spectrum, expDimRef.expDim)

else:
heavy_dim2 = findCcpnDataDim(spectrum, expDimRefX.expDim)
light_dim2 = findCcpnDataDim(spectrum, expDimRef.expDim)

else:

if light_dim1 is None:
heavy_dim1 = None
light_dim1 = findCcpnDataDim(spectrum, expDimRef.expDim)

else:
heavy_dim2 = None
light_dim2 = findCcpnDataDim(spectrum, expDimRef.expDim)


cross_peaks = []
for peak in peakList.peaks:
volume = peak.findFirstPeakIntensity(intensityType='volume')

if volume:
value = volume.value
err = volume.error
else:
value = None
err = None

volume = Datum(value, err)

height = peak.findFirstPeakIntensity(intensityType='height')
if height:
value = height.value
err = height.error
else:
value = None
err = None

intensity = Datum(value, err)

cross_peak = CrossPeak(number=peak.serial, volume=volume, intensity=intensity)

peakDims = []
for dim in (light_dim1, light_dim2, heavy_dim1 ,heavy_dim2):
peakDims.append(peak.findFirstPeakDim(dim=dim))


for i in range(len(ariaDimNames)):
peakDim = peakDims[i]
if peakDim:
setShiftFunc = getattr(cross_peak, 'set%sChemicalShift' % ariaDimNames[i])
setShiftFunc(ChemicalShift(peakDim.value, peakDim.valueError))

assignments = []
for peakDim in peakDims:
assignments0 = []

if peakDim:
for contrib in peakDim.peakDimContribs:
resonance = contrib.resonance
for ariaAtoms in getAriaAtomsFromResonance(resonance, ariaMolecule):
assi = Assignment(ariaAtoms, assignment_type=ASSIGNMENT_TYPE_MANUAL)
if peak.serial == 24:
print assi
print assignments0
if assi not in assignments0:
assignments0.append(Assignment(ariaAtoms, assignment_type=ASSIGNMENT_TYPE_MANUAL))

assignments.append(assignments0)

for i in range(len(ariaDimNames)):
addAssignFunc = getattr(cross_peak, 'add%sAssignment' % ariaDimNames[i])
for assignment in assignments[i]:
addAssignFunc(assignment)


cross_peaks.append(cross_peak)

spectrum = NOESYSpectrum(name=experiment.name, noes=cross_peaks)

return spectrum

def getAriaAtomsFromResonance(resonance, ariaMolecule, cache={}):
"""Descrn: Get the corresponding Aria2 atom in a molecule
given the input CCPN Resonance object
Inputs: Nmr.Resonance, Aria2 Molecule object
Output: List of List of Aria2 Atom objects
"""

from tools import string_to_segid

ariaAtoms = cache.get(resonance)
if ariaAtoms:
return ariaAtoms
else:
ariaAtoms = []

resonanceSet = resonance.resonanceSet
if not resonanceSet:
#print 'Attempt to get atoms from unassigned CCPN resonance (%d)' % resonance.serial
return []

atomSets = resonanceSet.sortedAtomSets()

residue = atomSets[0].findFirstAtom().residue

ariaChain = ariaMolecule.getChain(string_to_segid(residue.chain.code)) # Func warns if failure

ariaResidue = ariaChain.residues.get(residue.seqCode)
if not ariaResidue:
print 'Could not find ARIA Residue for CCPN residue %d%s' % (residue.seqCode,residue.ccpCode)
return []


#for atomSet in atomSets:
# - Removed this loop and added below.
# - Loop was giving two ARIA atoms for prochirals: now using only one
# because the restraint re-export was giving two resonances where
# it couldn't tell which was which: both linked to the same two atoms

# TJS Added
index = resonanceSet.sortedResonances().index(resonance)
atomSet = atomSets[min(index,len(atomSets)-1)]

atoms = []
for atom in atomSet.atoms:
ariaAtom = ariaResidue.atoms.get(atom.name)
if not ariaAtom:
print 'Could not find ARIA Atom for CCPN atom %d%s %s' % (residue.seqCode,residue.ccpCode, atom.name)
return []

atoms.append(ariaAtom)

ariaAtoms.append(atoms)

cache[resonance] = ariaAtoms

return ariaAtoms


def findCcpnDataDim(spectrum, expDim):
"""Descrn: Get the data dimension number that corresponds
to a given experimental dimension of a spectrum
Inputs: ccp.nmr.Nmr.DataSource, ccp.nmr.Nmr.ExpDim
Output: Int
"""

dataDim = None
if expDim:
dataDim = spectrum.findFirstDataDim(expDim=expDim)

if dataDim:
return dataDim.dim

# BARDIAUX
def dumpRestraintsList(constraintList, data_dir, type):
"""
Dump CCPN RestraintList to CNS Format TBL file.
"""

import DataContainer as DC
import os
from ccpnmr.format.converters.CnsFormat import CnsFormat

ccpProject = constraintList.root#nmrConstraintHead.project

cns_format = CnsFormat(ccpProject)

f = {DC.DATA_UNAMBIGUOUS : cns_format.writeDistanceConstraints,
DC.DATA_AMBIGUOUS : cns_format.writeDistanceConstraints,
DC.DATA_HBONDS : cns_format.writeHBondConstraints,
DC.DATA_DIHEDRALS : cns_format.writeDihedralConstraints,
DC.DATA_RDCS : cns_format.writeRdcConstraints,
DC.DATA_KARPLUS : cns_format.writeJCouplingConstraints,
DC.DATA_SSBONDS : cns_format.writeDistanceConstraints}


kw = {'compressResonances' : 0,
'minimalPrompts' : 1,
'forceNamingSystem' : 'XPLOR'}

fileName = getObjectKeyString(constraintList, delimiter='_').replace(" ", "")
fileName = "%s.tbl" % fileName

dst = os.path.join(data_dir, fileName)

# assumes that Resonnances are already linked
# overwrite existing file without warning

# attribute 'usePeakInfo' has default True now, but not all ccpn projects has
# this attribute and it's not necessary to ARIA 2 any longer
if type in (DC.DATA_AMBIGUOUS, DC.DATA_UNAMBIGUOUS): #by AWSS
kw['usePeakInfo'] = False #by AWSS

f[type](dst, constraintList = constraintList, **kw)

return dst

# BARDIAUX
def getAriaDistanceRestraintsList(constraint_list, constraint_type, aria_mol):
"""
Convert CCPN DistanceContraintList
to list of ARIA DistanceRestraint
"""

from ShiftAssignment import SpinSystem, AVERAGING_METHOD_NONE
import Contribution as C
from Singleton import SpinPairFactory
from AriaPeak import DistanceRestraint

# BARDIAUX 2.2 NEW
from CrossPeak import CrossPeak
from NOESYSpectrum import ConstraintList
from N import power
from DataContainer import PeakData#, LowerBoundCorrection, UpperBoundCorrection
from Datum import Datum

ccpn_project = constraint_list.root #nmrConstraintHead.project


SpinPair = SpinPairFactory()

restraints = []
peaks = []

#source = getObjectKeyString(constraint_list)
source = "%s %s" % (constraint_list.name, getObjectKeyString(constraint_list))

msg = "WARNING : %s ignored."

for distConstr in constraint_list.constraints:

spin_systems_pairs = []

# BARDIAUX 2.2 NEW
if distConstr.origData:
vol = distConstr.origData
else:
vol = power(distConstr.targetValue, -6)

vol = Datum(vol, 0)
xpk = CrossPeak(distConstr.serial, vol, vol)

contributions = []

restraint = DistanceRestraint()

target, upper, lower = distConstr.targetValue, distConstr.upperLimit, distConstr.lowerLimit
weight = distConstr.weight

restraint.setDistance(target)
restraint.setLowerBound(lower)
restraint.setUpperBound(upper)
restraint.setWeight(weight)


for constrItem in distConstr.items:

reso1, reso2 = constrItem.resonances

ss1 = SpinSystem(AVERAGING_METHOD_NONE)
atoms = getAriaAtomsFromResonance(reso1, aria_mol)

if not atoms:
print msg % (constrItem)
continue
elif len(atoms)> 1:
atoms = map(lambda x: x[0], atoms)
else:
atoms = atoms[0]

ss1.setAtoms(tuple(atoms))

ss2 = SpinSystem(AVERAGING_METHOD_NONE)
atoms = getAriaAtomsFromResonance(reso2, aria_mol)

if not atoms:
print msg % (constrItem)
continue
elif len(atoms)> 1:
atoms = map(lambda x: x[0], atoms)
else:
atoms = atoms[0]

ss2.setAtoms(tuple(atoms))

spin_pairs = []
for a1 in ss1.getAtoms():
for a2 in ss2.getAtoms():

if a1 == a2:
continue

sp = SpinPair(a1, a2)
spin_pairs.append(sp)

if not len(spin_pairs):
continue

if (ss1,ss2) in spin_systems_pairs or \
(ss2,ss1) in spin_systems_pairs :
continue

spin_systems_pairs.append((ss1,ss2))

c = C.Contribution(0,
C.CONTRIBUTION_TYPE_FAST_EXCHANGE,
spin_pairs, spin_systems = (ss1, ss2))



## TODO: hack
c.setSpinSystems((ss1, ss2))

contributions.append(c)

## Need a weight if no further Contribution evaluator
c.setWeight(1.)

if len(contributions) == 0:
print msg % (distConstr)
continue

restraint.setContributions(contributions)

peaks.append(xpk)
restraint.setReferencePeak(xpk)

restraints.append(restraint)

# assign a ARIA ConstraintList as Spectrum
s = ConstraintList(source, peaks)
pd = PeakData()
pd.reset()
pd['ccpn_id'] = getObjectKeyString(constraint_list)
s.setDataSource(pd)

return restraints, s
# return restraints

# BARDIAUX 2.2
def getCcpnExperimentData(peakList):

experiment = peakList.dataSource.experiment
frequency = mixing = None

# Spec freq
spec = experiment.spectrometer
if spec:
frequency = spec.protonFreq

# Mixing time
# CCPN return seconds, ARIA needs ms
transfer = experiment.findFirstExpTransfer(transferType='NOESY')
if transfer:
mixing = transfer.mixingTime
if mixing:
mixing = mixing *1e3

# Correlation Time ?
return frequency, mixing



## class CCPNData(Settings):

## def create(self):

## from Settings import MultiTypeEntity, AbsolutePath, TypeEntity
## from TypeChecking import TUPLE

## d = {}

## d['filename'] = AbsolutePath(exists=0)

## return d

## def create_default_values(self):

## d = {'filename': ''}

## return d

## class CCPNDataXMLPickler(XMLBasePickler):

## order = 'filename',

## def create(self):
## return CCPNData()

## def _xml_state(self, x):

## from xmlutils import XMLElement

## e = XMLElement()

## e.filename = x['filename']

## order = list(self.order)

## e.set_tag_order(order)

## return e

## def load_from_element(self, e):

## from tools import as_tuple

## s = self.create()

## filename = str(e.filename).strip()

## E = s.getEntity('filename')

## E.reset()
## E.mandatory(filename <> '')

## s['filename'] = filename

## return s

## CCPNData._xml_state = CCPNDataXMLPickler()._xml_state


"""

ARIA -- Ambiguous Restraints for Iterative Assignment

A software for automated NOE assignment

Version 2.2


Copyright (C) Benjamin Bardiaux, Michael Habeck, Therese Malliavin,
Wolfgang Rieping, and Michael Nilges

All rights reserved.


NO WARRANTY. This software package is provided 'as is' without warranty of
any kind, expressed or implied, including, but not limited to the implied
warranties of merchantability and fitness for a particular purpose or
a warranty of non-infringement.

Distribution of substantively modified versions of this module is
prohibited without the explicit permission of the copyright holders.

$Author: bardiaux $
$Revision: 1.14 $
$Date: 2007/11/26 14:23:53 $
"""



from aria import *
from Settings import Settings
from xmlutils import XMLElement, XMLBasePickler
from N import *
from time import clock
from time import ctime

## BARDIAUX 280105
REPORT_UPDATED_SPECTRUM = '.assigned'

REPORT_SUMMARY = 'report'
REPORT_NOE_RESTRAINTS = 'noe_restraints'
MISSING_STRUCTURES = 'missing_structures'
KEYBOARD_INTERRUPT = 'keyboard_interrupt'


## local path where MOLMOL restraints files are stored
PATH_MOLMOL = 'molmol'

def files_exist(file_list):
import os

for fn in file_list:
fn = os.path.expanduser(fn)
if not os.path.exists(fn):
return 0

return 1

def dump_peaks_as_text(peak_list, filename, gzip, options = None):

check_type(peak_list, LIST, TUPLE)
check_string(filename)
check_int(gzip)
check_type(options, NONE, DICT)

import AriaPeak

## write contributions as text file (both, ambiguous and
## unambiguous in one file)

## TODO: so far, text-pickler options are hard-coded.

settings = AriaPeak.AriaPeakListTextPicklerSettings()

## set options

if options is not None:
settings.update(options)

pickler = AriaPeak.AriaPeakListTextPickler(settings)

## sort peak list such, that active restraints come first,
## inactive last.

active = [p for p in peak_list if p.isActive()]
inactive = [p for p in peak_list if not p.isActive()]

peak_list = active + inactive

## Dump assignments

pickler.dump_assignments(peak_list, filename + '.assignments')

## decompose peak_list into ambiguous and unambiguous
## restraints

ambig = [p for p in peak_list if p.isAmbiguous()]
unambig = [p for p in peak_list if not p.isAmbiguous()]

pickler.dump_ambiguous(ambig, filename + '.ambig', gzip)
pickler.dump_unambiguous(unambig, filename + '.unambig', gzip)

## Dump restraint list, sorted by violations.
## Active restraints come first, inactive last

violated = [p for p in peak_list if p.analysis.isViolated()]

if violated:

d_lower = []
d_upper = []

lower = []
upper = []

for r in violated:

## In case of a lower-bound violation,
## put restraint in list of lower-bound-violating
## restraints

d = r.analysis.getLowerBoundViolation()[0]

if d is not None and d > 0.:
d_lower.append(d)
lower.append(r)

else:
d = r.analysis.getUpperBoundViolation()[0]

if d is not None and d > 0.:
d_upper.append(d)
upper.append(r)

lower_indices = argsort(d_lower).tolist()
lower_indices.reverse()

upper_indices = argsort(d_upper).tolist()
upper_indices.reverse()

a = [lower[i] for i in lower_indices]
b = [upper[i] for i in upper_indices]

viol = a + b

else:
viol = []

active = [p for p in viol if p.isActive()]
inactive = [p for p in viol if not p.isActive()]

viol = active + inactive

pickler.dump_violations(viol, filename + '.violations', gzip)

class IterationSettings(Settings):

def create(self):

from Settings import TypeEntity, ChoiceEntity, NonNegativeInt

kk = {}

kk['number'] = NonNegativeInt()
kk['number_of_structures'] = NonNegativeInt()
kk['number_of_best_structures'] = NonNegativeInt()

choices = ['total_energy', 'restraint_energy','noe_violations', 'restraint_violations'] # BARDIAUX 2.2
msg = 'Sort structures according to one of the following' + \
'choices: %s' % ', '.join(choices)
entity = ChoiceEntity(choices, error_message = msg)
kk['sort_criterion'] = entity

msg = 'Argument has to be a %s instance.'

d = {'calibrator_settings': 'CalibratorSettings',
'peak_assigner_settings': 'PeakAssignerSettings',
'violation_analyser_settings': 'ViolationAnalyserSettings',
'contribution_assigner_settings': 'ContributionAssignerSettings',
'merger_settings': 'MergerSettings',
'network_anchoring_settings' : 'NetworkSettings'}

for name, type in d.items():
entity = TypeEntity(type, error_message = msg % type)
kk[name] = entity

return kk

def create_default_values(self):

import Calibrator, PeakAssigner
import ViolationAnalyser as VA
import ContributionAssigner as CA
import Merger
## BARDIAUX 2.2
import Network

d = {}

d['calibrator_settings'] = Calibrator.CalibratorSettings()
d['peak_assigner_settings'] = PeakAssigner.PeakAssignerSettings()
d['violation_analyser_settings'] = VA.ViolationAnalyserSettings()
d['contribution_assigner_settings'] = CA.ContributionAssignerSettings()
d['merger_settings'] = Merger.MergerSettings()
## BARDIAUX 2.2
d['network_anchoring_settings'] = Network.NetworkSettings()

for value in d.values():
value.reset()

d['number'] = 0
d['number_of_structures'] = 20
d['number_of_best_structures'] = 7
d['sort_criterion'] = 'total_energy'

return d


class IterationSettingsXMLPickler(XMLBasePickler):

order = ['number', 'n_structures', 'sort_criterion',
'n_best_structures', 'assignment', 'merging',
'calibration', 'violation_analysis',
'partial_assignment', 'network_anchoring']

def _xml_state(self, x):

e = XMLElement(tag_order = self.order)

e.number = x['number']
e.n_structures = x['number_of_structures']
e.sort_criterion = x['sort_criterion']
e.n_best_structures = x['number_of_best_structures']

e.assignment = x['peak_assigner_settings']
e.merging = x['merger_settings']
e.calibration = x['calibrator_settings']
e.violation_analysis = x['violation_analyser_settings']
e.partial_assignment = x['contribution_assigner_settings']

# BARDIAUX 2.2
e.network_anchoring = x['network_anchoring_settings']

return e

def load_from_element(self, e):

s = IterationSettings()

s['number'] = int(e.number)
s['number_of_structures'] = int(e.n_structures)
s['sort_criterion'] = str(e.sort_criterion)
s['number_of_best_structures'] = int(e.n_best_structures)
s['peak_assigner_settings'] = e.assignment
s['merger_settings'] = e.merging
s['calibrator_settings'] = e.calibration
s['violation_analyser_settings'] = e.violation_analysis
s['contribution_assigner_settings'] = e.partial_assignment

## BARDIAUX 2.2
if hasattr(e, 'network_anchoring'):

s['network_anchoring_settings'] = e.network_anchoring
else:
import Network
nas = Network.NetworkSettings()
nas.reset()
s['network_anchoring_settings'] = nas


return s

class ProtocolSettings(Settings):

def create(self):

from Settings import TypeEntity, ChoiceEntity

d = {}

## water-refinement parameters

data_type = 'WaterRefinementParameters'
msg = 'Argument has to be a %s instance.' % data_type
entity = TypeEntity(data_type, error_message = msg)

d['water_refinement'] = entity

## iteration settings

entity = TypeEntity(DICT)
entity.set({})

d['iteration_settings'] = entity

## floating assignment flag

choices = [YES, NO]
msg = 'Floating assignment: %s?' % ' / '.join(choices)
entity = ChoiceEntity(choices, error_message = msg)

d['floating_assignment'] = entity

return d

def create_default_values(self):
return {'floating_assignment': YES}

def addIterationSettings(self, settings):

check_type(settings, 'IterationSettings')

number = settings['number']
it_settings = self['iteration_settings']

if not it_settings.has_key(number):
it_settings[number] = settings
else:
msg = 'IterationSettings instance with same number ' + \
'(%d) already stored in protocol settings.' % number
self.error(KeyError, msg)

def delIterationSettings(self, s):

check_type(s, 'IterationSettings')

number = s['number']
it_settings = self['iteration_settings']

if it_settings.has_key(number):
del it_settings[number]
else:
msg = 'IterationSettings (number %d) d not exist.' % number
self.error(KeyError, msg)

class Protocol(AriaBaseClass):

def __init__(self, settings):

check_type(settings, 'ProtocolSettings')

AriaBaseClass.__init__(self, name = 'Protocol')
self.setSettings(settings)

self.__reports_written = 0
self.constraintSetSerial = None

## self.iterations = {}

def getIterationSettings(self, number):

check_int(number)
return self.getSettings()['iteration_settings'][number]

def getInfrastructure(self):
from Singleton import ProjectSingleton

project = ProjectSingleton()
return project.getInfrastructure()

def initialize(self):
"""
must be called after unpickling a protocol in
order to finalize its initialization.
create objects which implement the actual aria-method
set parameters of first iteration.
"""

from NOEModel import NOEModel, ISPA
import Calibrator
import ViolationAnalyser as VA
import ContributionAssigner as CA
import Merger
## BARDIAUX 2.2
import Network

## NOE model
## TODO: not nessecary to store model!

#self.model = NOEModel()
self.model = ISPA()

## calibrator

cs = Calibrator.CalibratorSettings()
self.calibrator = Calibrator.Calibrator(cs, self.model)

## violation-analyser

vas = VA.ViolationAnalyserSettings()
self.violation_analyser = VA.ViolationAnalyser(vas)

## contribution assigner

cas = CA.ContributionAssignerSettings()
self.contribution_assigner = CA.ContributionAssigner(cas)

## merger

ms = Merger.MergerSettings()
self.merger = Merger.Merger(ms)

## BARDIAUX 2.2
## network
nas = Network.NetworkSettings()
self.network_anchoring = Network.NetworkAnchoring(nas)

def _updateIterationSettings(self, iteration):
"""
applies the parameters-settings in 'iteration' to
sub-modules
"""

check_type(iteration, 'Iteration')

settings = self.getIterationSettings(iteration.getNumber())

cs = settings['calibrator_settings']
self.calibrator.getSettings().update(cs)

vas = settings['violation_analyser_settings']
self.violation_analyser.getSettings().update(vas)

cas = settings['contribution_assigner_settings']
self.contribution_assigner.getSettings().update(cas)

ms = settings['merger_settings']
self.merger.getSettings().update(ms)

## BARDIAUX 2.2
nas = settings['network_anchoring_settings']
self.network_anchoring.getSettings().update(nas)


def _updateSpectrumSettings(self, spectrum):
spectrum_data = spectrum.getDataSource()
self.calibrator.getSettings().update(spectrum_data)
self.violation_analyser.getSettings().update(spectrum_data)

def mergePeakLists(self, iteration):

## BARDIAUX 2.2
# remove previous combined peaks
iteration.setCombinedRestraints([])

if self.merger.getSettings()['method'] == 'no_merging':
return

peaks = [p for p in iteration.getPeakList() if p.isActive()]

## merge peaks. peaks which are merged (away), i.e.
## are marked via peak.isMerged(1)

n, combined = self.merger(peaks)

# add combined restraints
iteration.setCombinedRestraints(combined)

if n:
msg = 'Peaks merged: %d / %d peaks (%.1f %%) (see report for details).'
self.message(msg % (n, len(peaks), 100. * n / len(peaks)))

if combined:
msg = 'Peaks combined: %d / %d peaks (%.1f %%).'
self.message(msg % (len(combined), len(peaks), 100. * len(combined) / len(peaks)))

## self.message('old %d new %d.' % (len(peaks), len([p for p in iteration.getPeakList() if p.isActive()])))


def compileFileList(self, number, **kw):
"""
Compiles a tuple of filenames which must exist after
iteration 'number' has been successfully completed.
If the method is called with keywords, the returned tuple
contains only a sub-set of all filenames. The sub-set depends
on the particular keyword(s)
"""

check_int(number)

STRUCTURES = 'structures'
FLOAT_FILES = 'float_files'

from os.path import join

keywords = (STRUCTURES, FLOAT_FILES)

unknown = [key for key in kw if key not in keywords]

if unknown:
s = 'Unknown sub-list identifier ("%s"). Known identifier ' + \
'are: %s'
self.error(s % (str(unknown), str(keywords)))

if kw:
keywords = kw.keys()

infra = self.getInfrastructure()

## structures (pdb-files)

it_settings = self.getIterationSettings(number)
n_structures = it_settings['number_of_structures']

path = infra.get_iteration_path(number)

l = []

if STRUCTURES in keywords:
name_template = infra.get_file_root() + '_%d.pdb'
l += [join(path, name_template % j) \
for j in range(1, n_structures + 1)]

if FLOAT_FILES in keywords:
name_template = infra.get_file_root() + '_%d.float'
l += [join(path, name_template % j) \
for j in range(1, n_structures + 1)]

return tuple(l)

def findFirstIteration(self):
"""
For every iteration, this method checks if the required
PDB-files exist. If so, it seeks to the next iteration until
it ends up at an iteration with missing files. If files are
present for all iterations, None is returned. Otherwise the
number of the iteration with missing files is returned.
"""

infra = self.getInfrastructure()
n_iterations = infra.getSettings()['n_iterations']

for i in range(n_iterations):

## Get filenames of all structures for iteration i

pdb_files = self.compileFileList(i, structures=1)
if not files_exist(pdb_files):
return i

return None

def _writePDBFileList(self, iteration):
import os

infra = self.getInfrastructure()

iteration_path = infra.get_iteration_path(iteration.getNumber())
molmol_path = os.path.join(iteration_path, PATH_MOLMOL)

ok = 1

if not os.path.exists(molmol_path):

try:
os.makedirs(molmol_path)

except Exception, msg:
import tools
self.warning(tools.last_traceback())
msg = 'Could not create directory for MOLMOL output.'
self.warning(msg)
ok = 0

if not ok:
return None

## file.nam

## files are sorted according to 'sort_criterion'
## which is part of the <structure_generation>
## section of project xml.

ensemble = iteration.getStructureEnsemble()
file_list = ensemble.getFiles()

if file_list is None:
s = 'Could not write MolMol file.nam: No PDB-files' + \
'in structure ensemble.'
self.warning(s)
return None

filename = os.path.join(molmol_path, 'file.nam')

try:
f = open(filename, 'w')
except:
self.warning('Could not create %s.' % filename)
return None

s = '\n'.join(file_list)

try:
f.write(s)
f.close()

except:
self.warning('Could not write PDB-file list (%s).' % filename)
return None

return molmol_path

def structure_calc_done(self, engine, molecule, iteration):

## Check if there are missing structures

if engine.missingStructures():
self.__done = MISSING_STRUCTURES
return

self.message('Structure calculation done.')

it_settings = self.getIterationSettings(iteration.getNumber())
ensemble = engine.getEnsemble(it_settings, molecule)

## store structure-ensemble in iteration

iteration.setStructureEnsemble(ensemble)

self.__done = 1

def startStructureCalculation(self, iteration, molecule):

from Singleton import ProjectSingleton

## get list of all restraints

peaks = iteration.getPeakList()


## BARDIAUX 2.2 Add user-distance CCPN restraints
restraints = iteration.getDistanceRestraintsList()
peaks += restraints
# add also combined restraints
combined = iteration.getCombinedRestraints()
peaks += combined

project = ProjectSingleton()
engine = project.getStructureEngine()

## callback that will be called when pdb-files have been generated.

f = lambda engine = engine, molecule = molecule, it = iteration, \
c = self.structure_calc_done: c(engine, molecule, it)

engine.set_callback(f)
engine.go(peaks, self.getSettings(), iteration.getNumber())

def doSolventRefinement(self, iteration, molecule):

from Singleton import ProjectSingleton
import StructureEnsemble as SE

## If the given iteration lacks a structure ensemble,
## load PDB-files from last iteration.

ensemble = iteration.getStructureEnsemble()

if ensemble is None:

## create structure-ensemble from given 'filenames'
number = iteration.getNumber()
it_settings = self.getIterationSettings(number)
se_settings = SE.StructureEnsembleSettings()

## Set iteration-specific settings

se_settings.update(it_settings)

ensemble = SE.StructureEnsemble(se_settings)

filenames = self.compileFileList(number, structures=1)
naming_convention = 'cns'
ensemble.read(filenames, molecule, naming_convention)

iteration.setStructureEnsemble(ensemble)

project = ProjectSingleton()
engine = project.getStructureEngine()

file_list = engine.solvent_refine(self.getSettings(), ensemble)

## cleanup (for solvent refinement and last iteration)

infra = self.getInfrastructure()
last_it = infra.getSettings()['n_iterations'] - 1

engine.cleanup(infra.get_refinement_path())
engine.cleanup(infra.get_iteration_path(last_it))

def doViolationAnalysis(self, restraints, ensemble, store_analysis = 0):
"""
'restraints': list of AriaPeaks. The bounds of every
restraint in that list is checked against distances found
in the 'ensemble'.

'targets': list of AriaPeaks. The violationAnalyser will
store all intermediate results in their analysis-section.
Note: we assume, that peaks[i] corresponds to results[i]
for all i !. If a restraint has been violated, the
corresponding 'target'_restraint will be marked as violated.
"""

f = self.violation_analyser.analysePeak

violated = []
non_violated = []

## get theshold for current iteration
settings = self.violation_analyser.getSettings()
threshold = settings['violation_threshold']

for restraint in restraints:

R_viol = f(ensemble, restraint, store_analysis)

##
## If a restraint has been violated in too many structures
## (according to 'threshold'), mark is a violated.
##

if R_viol > threshold:
restraint.analysis.isViolated(1)
violated.append(restraint)

else:
restraint.analysis.isViolated(0)
non_violated.append(restraint)

## For violated restraints: if bound-correction is enabled,
## repeat violation-analysis with modified bounds.

if settings['lower_bound_correction']['enabled'] == YES:
new_lower = settings['lower_bound_correction']['value']
else:
new_lower = None

if settings['upper_bound_correction']['enabled'] == YES:
new_upper = settings['upper_bound_correction']['value']
else:
new_upper = None

if new_lower is not None or new_upper is not None:

## We forget 'store_analysis' here, since it has already
## been stored (if set).

R_viol = [f(ensemble, r, lower_correction = new_lower,
upper_correction = new_upper) for r in violated]

## List of restraint-indices which are no longer
## violated after bound modification.

indices = nonzero(less(R_viol, threshold))
new_non_violated = [violated[i] for i in indices]

[r.analysis.isViolated(0) for r in new_non_violated]

else:
new_non_violated = None

return violated, non_violated, new_non_violated

def done(self):
self.message('finished.')

def _dump_peak_list(self, peaks, path):

check_list(peaks)
check_string(path)

import os
from Singleton import ProjectSingleton
import Merger

project = ProjectSingleton()
s = project.getReporter()['noe_restraint_list']

## Dump only non-merged restraitns

not_merged = [r for r in peaks if not r.isMerged()]

if s['xml_output'] in (YES, GZIP):

from AriaXML import AriaXMLPickler as pickler

filename = os.path.join(path, REPORT_NOE_RESTRAINTS + '.xml')
t = clock()
p = pickler()

# BARDIAUX 2.2: It's not possible to dump
# CCPN restraints as XML
not_merged = [r for r in not_merged if not is_type(r, 'DistanceRestraint')]

gzip = {YES: 0, GZIP: 1}[s['xml_output']]
p.dump(not_merged, filename, gzip = gzip)
self.message('NOE-restraint list (xml) written (%s).' % filename)
self.debug('Time: %ss' % str(clock()-t))

if s['text_output'] in (YES, GZIP):
t = clock()
filename = os.path.join(path, REPORT_NOE_RESTRAINTS)

gzip = {YES: 0, GZIP: 1}[s['text_output']]
dump_peaks_as_text(not_merged, filename, gzip)
self.message('NOE-restaint list (text) written (%s).' % filename)
self.debug('Time: %ss' % str(clock()-t))

if s['pickle_output'] in (YES, GZIP):

from tools import Dump

filename = os.path.join(path, REPORT_NOE_RESTRAINTS + '.pickle')
t = clock()

gzip = {YES: 0, GZIP: 1}[s['pickle_output']]
Dump(not_merged, filename, gzip = gzip)
self.message('NOE-restraint list (python pickle) ' + \
'written (%s).' % filename)
self.debug('Time: %ss' % str(clock()-t))

## Write Merger report

t = clock()

merged = [r for r in peaks if r.isMerged()]
filename = os.path.join(path, REPORT_NOE_RESTRAINTS + '.merged')

p = Merger.MergerTextPickler()
p.dump(merged, filename)
self.message('Report for Merging-step written (%s).' % filename)
self.debug('Time: %ss' % str(clock() - t))


def _dump_spectra(self, peaks, spectrum, filename):

check_list(peaks)
check_type(spectrum, 'NOESYSpectrum')
check_string(filename)

import os
from Singleton import ProjectSingleton
from NOESYSpectrum import NOESYSpectrum
import Assignment
from copy import deepcopy
from AriaXML import AriaXMLPickler as pickler

project = ProjectSingleton()

s = project.getReporter()['spectra']


if s['write_unambiguous_only'] == YES:
peaks = [pk for pk in peaks if not pk.isAmbiguous()]

if peaks and s['write_assigned'] == YES:

spectrum = deepcopy(spectrum)

for pk in peaks:

ref_pk = spectrum.findPeak(pk.getReferencePeak().getNumber())

if ref_pk is None:
self.error('Inconsistency: Could not find reference peak %d in copy of spectrum "%s"' %
(ref_pk.getNumber(), spectrum.getName()))

active_contribs = [c for c in pk.getContributions() \
if c.getWeight() > 0.]

if not ref_pk.isAssigned() or \
(ref_pk.isAssigned() and s['write_assigned_force'] == YES) and\
active_contribs:

h = active_contribs[0]

ss = h.getSpinSystems()

for sp_sys in ss:

dim = ref_pk.getDimension(sp_sys)

AUTO = Assignment.ASSIGNMENT_TYPE_AUTOMATIC

A = Assignment.Assignment(sp_sys.getAtoms(), AUTO)

if dim == 1:
ref_pk.setProton1Assignments((A,))
elif dim == 2:
ref_pk.setProton2Assignments((A,))

for h in active_contribs[1:]:

ss = h.getSpinSystems()

for sp_sys in ss:

dim = ref_pk.getDimension(sp_sys)

AUTO = Assignment.ASSIGNMENT_TYPE_AUTOMATIC

A = Assignment.Assignment(sp_sys.getAtoms(), AUTO)

if dim == 1:
atoms_assi = [a.getAtoms() for a in ref_pk.getProton1Assignments()]
if not A.getAtoms() in atoms_assi:
ref_pk.addProton1Assignment(A)
elif dim == 2:
atoms_assi = [a.getAtoms() for a in ref_pk.getProton2Assignments()]
if not A.getAtoms() in atoms_assi:
ref_pk.addProton2Assignment(A)


t = clock()
p = pickler()

p.dump(spectrum, filename, gzip = 0)
self.debug('Time: %ss' % str(clock()-t))

self.message('Assigned spectrum "%s" written to file %s"' \
% (spectrum.getName(), filename))

def _dump_ccpn(self, iteration, path, is_water_refinement=False):

from Singleton import ProjectSingleton

project = ProjectSingleton()

if project.ccpn_project_instance is None:
return

import exportToCcpn as ccpnExport

project_settings = project.getSettings()
run_name = project_settings['run'].strip().replace(' ', '_')
file_root = project_settings['file_root'].strip().replace(' ', '_')

ccpn_project = project.ccpn_project_instance
nmrProject = ccpn_project.currentNmrProject or ccpn_project.findFirstNmrProject()

## TBD: proper handling of NMR projects, e.g. the case in which
## multiple NMR projects exist.

if not nmrProject:
self.message('CCPN export: No NMR project found, creating new one.')

name = 'ARIA2_run_%s' % run_name

nmrProject = ccpn_project.newNmrProject(name=name)

s = project.getReporter()['ccpn']

export = 0

restraints = [r for r in iteration.getPeakList() if r.isActive()]

## TBD: multiple chains

aria_chain = project.getMolecule().get_chains()[0]

if s['export_structures'] == YES and \
(iteration.getNumber() == project_settings['n_iterations']-1 or \
is_water_refinement):

export = 1

# BARDIAUX: Update for multiple chains
chains = ccpnExport.getChainsFromAria2(restraints, ccpn_project,
aria_chain=aria_chain)

if not chains:
self.warning(StandardError, 'CCPN export: No molecular system found.')
return

structures = ccpnExport.getStructuresFromAria2Dir(path, chains[0])

if not structures:
self.warning('CCPN export: Unable to load any structures from iteration directory %s' % path)

else:

if self.constraintSetSerial is not None:
constraintSet = nmrProject.findFirstNmrConstraintStore(serial=self.constraintSetSerial)
self.message('CCPN export: Using existing constraint set.')

else:
self.message('CCPN export: Creating new constraint set.')
constraintSet = ccpnExport.makeNmrConstraintStore(nmrProject)
self.constraintSetSerial = constraintSet.serial

self.message('CCPN export: PDB files exported.')

struct_gen = ccpnExport.makeStructureGeneration(structures, constraintSet)

if not is_water_refinement:
struct_gen.name = 'ARIA2_%s_run_%s_iteration_%d' % (file_root, run_name, iteration.getNumber())
struct_gen.details = 'Structures created by ARIA2, run "%s", iteration %d.' % \
(project.getSettings()['run'], iteration.getNumber())

else:
struct_gen.name = 'ARIA2_%s_run_%s_water_refinement' % (file_root, run_name)
struct_gen.details = 'Structures created by ARIA2, run "%s", water refinement.'

else:
structures = None

if not is_water_refinement and (s['export_noe_restraint_list'] == 'all' or \
(s['export_noe_restraint_list'] == 'last' and \
iteration.getNumber() == project.getSettings()['n_iterations']-1)):

export = 1

if self.constraintSetSerial is not None:
constraintSet = nmrProject.findFirstNmrConstraintStore(serial=self.constraintSetSerial)
self.message('CCPN export: Using existing constraint set.')

else:
self.message('CCPN export: Creating new constraint set.')
constraintSet = ccpnExport.makeNmrConstraintStore(nmrProject)
self.constraintSetSerial = constraintSet.serial

# BARDIAUX: Update for multiple chains
chains = ccpnExport.getChainsFromAria2(restraints, ccpn_project,
aria_chain=aria_chain)

if restraints:
constraints = ccpnExport.getConstraintsFromAria2(restraints, chains, constraintSet,
structures)

constraints[0].name = 'ARIA2_%s_run%s_NOEs_it%d' % (file_root, run_name, iteration.getNumber())
constraints[0].details = 'ARIA2 NOE restraint list, project "%s", run "%s", iteration %d.' % \
(project.getSettings()['name'], project.getSettings()['run'], iteration.getNumber())

self.message('CCPN export: NOE restraint list exported.')

# BARDIAUX export Constraints from CCPN
for orig_list, distance_constraints in iteration.getDistanceRestraints().items():

distance_constraints = [r for r in distance_constraints if r.isActive()]

constraints = ccpnExport.getConstraintsFromAria2(distance_constraints, chains, constraintSet,
structures)

list_id = orig_list.getDataSource().get('ccpn_id')

constraints[0].name = 'ARIA2_%s_run%s_Dists_%s_it%d' % (file_root, run_name, list_id, iteration.getNumber())
constraints[0].details = 'ARIA2 DistanceConstraints list %s, project "%s", run "%s", iteration %d.' % \
(orig_list.getName(), project.getSettings()['name'], project.getSettings()['run'], iteration.getNumber())

self.message('CCPN export: DistanceConstraints list exported.')

if not is_water_refinement and (s['export_assignments'] == YES and \
iteration.getNumber() == project.getSettings()['n_iterations']-1):

import DataContainer as DC

export = 1

## compile dict that maps peaklist keys to name of new peaklist

ccpn_spectra = [x for x in project.ccpn_data_sources.get(DC.DATA_SPECTRUM, ())]

names_dict = {}

template = 'ARIA2_NOE_assignments_%s_run_%s_it_%d_%s'

for x in ccpn_spectra:

ccpn_id = x['peaks']['ccpn_id']

args = file_root, run_name, iteration.getNumber(), ccpn_id
names_dict[ccpn_id] = template % args

peak_list_keys = ccpnExport.getPeakAssignmentsFromAria2(ccpn_project, restraints, namesDict=names_dict,
aria_chain=aria_chain)

self.message('CCPN export: NOE assignments exported.')

if export:

try:
ccpn_project.saveModified()

except Exception, msg:

self.warning('CCPN export: Could not save project. Error message was: %s' % msg) #by AWSS, msg is an object not a string

## unload objects to free some memory

if structures:
for s in structures:
#by AWSS begin
# CCPN API changed the way how unload() is called
try:
s.unload()
except Exception, msg:
self.warning("Couldn't unload structure: %s" % msg)
#by AWSS end

if constraintSet:
#by AWSS begin
# CCPN API changed the way how unload() is called
try:
constraintSet.unload()
except Exception, msg:
self.warning("Couldn't unload constraintSet: %s" % msg)
#by AWSS end

self.message('CCPN project saved.')

def _dump_iteration(self, iteration, path):

import os, Iteration

filename = os.path.join(path, REPORT_SUMMARY)
pickler = Iteration.IterationTextPickler()
pickler.dump(iteration, filename)

## BARDIAUX 2.2
def _dump_rms(self, peaks, iteration_number):

import RmsReport
infra = self.getInfrastructure()
text_path = infra.get_iteration_path(iteration_number)
graphics_path = infra.get_iteration_graphics_path(iteration_number)

rp = RmsReport.RmsReport(peaks, iteration_number, text_path, graphics_path)
rp.go()


def dumpIteration(self, iteration):

import os
from Singleton import ProjectSingleton

project = ProjectSingleton()
check_type(iteration, 'Iteration')

infra = self.getInfrastructure()
path = infra.get_iteration_path(iteration.getNumber())

# BARDIAUX 2.2
# dump also CCPN DistanceRestraints
# all_peaks = iteration.getPeaks()
all_peaks = {}
all_peaks.update(iteration.getPeaks())
all_peaks.update(iteration.getDistanceRestraints())

order = all_peaks.keys()
order.sort()

peak_list = []

for spectrum in order:

peaks = all_peaks[spectrum]

## sort peak_list wrt ref_peak number

peaks.sort(lambda a, b, c = cmp: \
c(a.getReferencePeak().getNumber(), \
b.getReferencePeak().getNumber()))


peak_list += peaks

## BARDIAUX 25/04/05
# Dump updated spectrum

if is_type(spectrum, 'ConstraintList'):
# Not for CCPN ConstraintList
continue

spec_name = spectrum.getName()
spec_name = spec_name.replace(' ','_')

filename = os.path.join(path, spec_name + \
REPORT_UPDATED_SPECTRUM +'.xml')


s = project.getReporter()['spectra']

if s['iteration'] == 'all' or \
(s['iteration'] == 'last' and \
iteration.getNumber() == project.getSettings()['n_iterations']-1):

self._dump_spectra(peaks, spectrum, filename)


self._dump_peak_list(peak_list, path)
self._dump_iteration(iteration, path)
self._dump_ccpn(iteration, path)

## BARDIAUX 2.2
## RMS report

if peak_list and iteration.getNumber() > 0:

try:
self._dump_rms(peak_list, iteration.getNumber())
except:
self.message('Error during RMS analysis/graphics generation.')


self.__reports_written = 1

def reportsWritten(self):
"""
returns 1 if any report files have been written.
"""

return self.__reports_written

def _get_peak_sizes(self, peaks):
"""
depending on the calibrator-setting 'volume_or_intensity',
the method returns volumes or intensities
"""

ref_peaks = [p.getReferencePeak() for p in peaks]

if self.calibrator.getSettings()['volume_or_intensity'] == 'volume':
peak_sizes = [p.getVolume()[0] for p in ref_peaks]
else:
peak_sizes = [p.getIntensity()[0] for p in ref_peaks]

return peak_sizes

# Bardiaux rMat
def _get_peak_theoric_volumes(self, peaks):

"""
get the theoric volume from the spin diffusin correction.
"""
return array([p.getTheoricVolume() for p in peaks])

# BARDIAUX 2.2
# get settings of DistanceRestraint sourec list
def __getListSource(self, p):
return p.getReferencePeak().getSpectrum().getListSource()

def setModelIntensities(self, restraints, ensemble, calibration_factor):

from Datum import Datum

f = self.model.calculatePeaksize

model_peak_sizes = array([f(r, ensemble) for r in restraints])
calculated_peak_sizes = model_peak_sizes * calibration_factor

for i in range(len(restraints)):
d = Datum(calculated_peak_sizes[i], None)
restraints[i].analysis.setCalculatedPeaksize(d)

def calculateBounds(self, factor, peaks, bound_corrected = None, ensemble = None):
"""
calculate lower- and upper bounds for every peak using
the calibration 'factor'. values are stored.
'bound_corrected': list of restraints which are classified
as correct after bound-modification.
"""
# BARDIAUX 2.2
# ConstraintList can bypass the calibration
if is_type(peaks[0], 'DistanceRestraint'):
calibrate = self.__getListSource(peaks[0])['calibrate']
if calibrate == NO or \
(self.findFirstIteration() == 0 and calibrate == 'all_iterations_except_first'):
return

factor = power(factor, 1./6)

peak_sizes = self._get_peak_sizes(peaks)

# Malliavin/Bardiaux rMat
cs = self.calibrator.getSettings()
if cs['relaxation_matrix'] == YES and ensemble is not None:

## from NOEModel import ISPA
## ispa = ISPA()
## f = ispa.calculatePeaksize

## ispa_peak_sizes = array([f(r, ensemble) for r in peaks])

ispa_peak_sizes = array([p.getIspa() for p in peaks])
peak_theoric_vol = self._get_peak_theoric_volumes(peaks)

ratio = ispa_peak_sizes / peak_theoric_vol
distances = factor * power(peak_sizes * ratio, -1. / 6)


else:

distances = factor * power(peak_sizes, -1. / 6)

## TODO: hard-coded 0.125

if cs['error_estimator'] == 'intensity':
errors = 0.125 * power((factor * power(peak_sizes, -1. / 6)), 2.)

else:
errors = 0.125 * power(distances, 2.)

## # BARDIAUX : To be implemented
## if cs['error_estimator'] == 'no_errors':
## errors = [0.] * len(distances)

# errors = [0.] * len(distances)

## distances = factor * power(peak_sizes, -1. / 6)
## ## TODO: hard-coded 0.125
## errors = 0.125 * power(distances, 2.)

## lower bounds are >= 0.

lower_bounds = clip(distances - errors, 0., 1.e10)
upper_bounds = distances + errors

for i in range(len(peaks)):

peak = peaks[i]

peak.setDistance(distances[i])
peak.setLowerBound(lower_bounds[i])
peak.setUpperBound(upper_bounds[i])

## Set new (fixed) bounds for bound-corrected restraints

if bound_corrected:
va_settings = self.violation_analyser.getSettings()

if va_settings['lower_bound_correction']['enabled'] == YES:
new_bound = va_settings['lower_bound_correction']['value']
[r.setLowerBound(new_bound) for r in bound_corrected]

if va_settings['upper_bound_correction']['enabled'] == YES:
new_bound = va_settings['upper_bound_correction']['value']
[r.setUpperBound(new_bound) for r in bound_corrected]

def doDumboCalibration(self, peaks):

# BARDIAUX 2.2
# ConstraintList can bypass calibration
if is_type(peaks[0], 'DistanceRestraint'):
do_cal = self.__getListSource(peaks[0])['calibrate'] == 'all_iterations_except_first'
if do_cal:
return 1.

peak_sizes = self._get_peak_sizes(peaks)

## Assume that average distance of atoms
## causing an NOE is 3.0A

d_calib = 3.0

sum_noe_calc = len(peaks) * (d_calib ** -6)

factor = sum(peak_sizes) / sum_noe_calc

return factor

def doCalibration(self, restraints, ensemble, store_analysis = 0):

do_cal = 1

# BARDIAUX 2.2
# ConstraintList can bypass calibration

if is_type(restraints[0], 'DistanceRestraint'):
do_cal = self.__getListSource(restraints[0])['calibrate'] == NO
if do_cal:
return 1.

if ensemble is None:
factor = self.doDumboCalibration(restraints)
else:
f = self.calibrator.calculateEstimator
factor = f(restraints, ensemble, store_analysis = store_analysis)

## if factor is None, i.e. if no NOEs stronger than a
## certain cutoff were found, set the cutoff to zero
## and calibrate again.

if factor is None:

s = self.calibrator.getSettings()
d_cutoff = s['distance_cutoff']

s = 'Could not perform 1st calibration, since ' + \
'no distances less than %.1f A were found in the ' + \
'ensemble. Omitting distance-cutoff and ' + \
'calibrating again...'

self.message(s % d_cutoff)

factor = f(restraints, ensemble, \
store_analysis = store_analysis, use_cutoff = 0)

return factor

def finalize_engine(self, molecule):
from Singleton import ProjectSingleton

project = ProjectSingleton()
engine = project.getStructureEngine()
engine.prepare(self.getSettings(), molecule)

def doAnalysis(self, restraints, ensemble):

##
## 1st calibration
##
## Store results.
##

factor = self.doCalibration(restraints, ensemble, store_analysis = 1)

##
## Violation analysis
##
## Store results:
##
## - average violation distance
## - whether a restraint is violated
## - degree of violation etc.
##

violated, non_violated, new_non_violated = \
self.doViolationAnalysis(restraints, ensemble,
store_analysis = 1)

## If possible, enlarge set of non-violated restraints

if new_non_violated:
non_violated += new_non_violated

##
## 2nd calibration
##

if non_violated:
factor = self.doCalibration(non_violated, ensemble)

## Calculate model peak-sizes

self.setModelIntensities(restraints, ensemble, factor)

def export_water_refined_structures(self, iteration):

if self.getSettings()['water_refinement']['enabled'] == NO:
return

infra = self.getInfrastructure()
path = infra.get_refinement_path()

self._dump_ccpn(iteration, path, is_water_refinement=True)

def writeReports(self, iteration):

t = clock()

s = 'Performing analysis on calculated structures...'
self.message(s)

ensemble = iteration.getStructureEnsemble()

# BARDIAUX 2.2
# add CCPN DistanceRestraints
all_restraints = {}
all_restraints.update(iteration.getPeaks())
all_restraints.update(iteration.getDistanceRestraints())

for spectrum, restraints in all_restraints.items():
#for spectrum, restraints in iteration.getPeaks().items():

## Set spectrum-specific settings
self._updateSpectrumSettings(spectrum)
self.doAnalysis(restraints, ensemble)

self.message('Analysis done.')
self.debug('Time: %ss' % str(clock() - t))

self.dumpIteration(iteration)

def readPDBFiles(self, filenames, molecule, float_files = None,
format = 'cns'):
"""
'filenames' is a dict or a tuple. In case of a dict,
keys are filenames, values are format-strings. In case of
a tuple, the argument 'format' is used as format-string.
Reads PDB-files 'filenames' and the respective
'float_files'. Returns a StructureEnsemble.
"""

check_type(filenames, TUPLE, DICT)
check_type(molecule, 'Molecule')
check_type(float_files, TUPLE, NONE)
check_string(format)

import StructureEnsemble as SE

if float_files == ():
float_files = None

if float_files is not None:

from FloatFile import FloatFile

parser = FloatFile()
swapped_atoms = [parser.parse(f) for f in float_files]

else:
swapped_atoms = None

## create structure-ensemble from given 'filenames'

se_settings = SE.StructureEnsembleSettings()
ensemble = SE.StructureEnsemble(se_settings)
ensemble.read(filenames, molecule, format, swapped_atoms)

return ensemble

def start(self, iteration, molecule):
"""
setup the first iteration. calculate distance-estimates
using either a template structure or by doing a dumbo-
calibration.
"""

check_type(molecule, 'Molecule')
check_type(iteration, 'Iteration')

## set-up first iteration

import os
from Singleton import ProjectSingleton
from DataContainer import DATA_TEMPLATE_STRUCTURE

t_iteration = clock()

## create seed-assignment

## if a template-structure has been specified,
## use that structure to create a better seed-assignment.
## otherwise we use an extended chain (in case of CNS,
## it is created by using generate_template.inp

infra = self.getInfrastructure()
project = ProjectSingleton()

templates = project.getData(DATA_TEMPLATE_STRUCTURE)
templates = [t for t in templates if t['enabled'] == YES]

## If template structures have been specified, we
## create an initial structure ensemble which will
## be used in the following steps to calculate
## an initial calibration factor etc.

if templates:

filenames = [t['filename'] for t in templates]
formats = [t['format'] for t in templates]

s = 'Template structure(s) specified (%s). Using template(s)' + \
' for calibration / seed-assignment generation.'

self.message(s % ', '.join(map(os.path.basename, filenames)))

## Compile local filenames (template PDB-files are
## read from local data-directory)

filenames = [infra.get_local_filename(f, DATA_TEMPLATE_STRUCTURE) \
for f in filenames]

## Build dict with filenames as keys and formats as
## values.

d = {}

for i in range(len(formats)):
d[filenames[i]] = formats[i]

ensemble = self.readPDBFiles(d, molecule)
ensemble.getSettings()['number_of_best_structures'] = 'all'
iteration.setStructureEnsemble(ensemble)

self.message('Initial iteration created.', verbose_level = VL_LOW)
self.debug('Time: %ss' % str(clock() -t_iteration))

## register initial iteration and start protocol

## self.iterations[iteration.getNumber()] = iteration

## Return last iteration
return self.run_protocol(molecule, iteration)

def go(self, iteration, molecule):

check_type(molecule, 'Molecule')
check_type(iteration, 'Iteration')

## Check which iterations have already been
## calculated, so that we can skip them.

first_iteration = self.findFirstIteration()

## If no iteration has been calculated yet, run the
## full protocol.

if first_iteration == 0:
iteration = self.start(iteration, molecule)

## If some iterations already exist, calculate the remaing ones.

elif first_iteration is not None:

t = clock()

s = 'Existing iterations found. Resuming with iteration ' + \
'%d.'

self.message(s % first_iteration)

iteration._setNumber(first_iteration - 1)

## Get filenames of structures which ought exist
## after parent-iteration has been successfully completed.

CFL = self.compileFileList

filenames = CFL(first_iteration - 1, structures = 1)
float_files = CFL(first_iteration - 1, float_files = 1)

it_settings = self.getIterationSettings(iteration.getNumber())

## Default PDB-format is 'cns'

ensemble = self.readPDBFiles(filenames, molecule,
float_files)


## Apply structure-ensemble settings of current
## iteration.

se_settings = ensemble.getSettings()
se_settings.update(it_settings)

ensemble.settingsChanged()

iteration.setStructureEnsemble(ensemble)

s = 'PDB-files / FLOAT-files for iteration %d read: %ss'
self.debug(s % (first_iteration, str(clock() - t)))

iteration = self.run_protocol(molecule, iteration)

else:

from Iteration import Iteration

self.message('Iterations are already complete.')

## TODO: this is a bit fiddled!
## Since the solvent-refinement step does not refer to any
## restraint, we simply create an empty iteration
## and set its number to the last iteration's.

infra = self.getInfrastructure()
last_iteration = infra.getSettings()['n_iterations'] - 1
iteration = Iteration(last_iteration)

## Start water refinement
self.doSolventRefinement(iteration, molecule)

## Export structures to CCPN

self.export_water_refined_structures(iteration)

## End of ARIA protocol.
self.done()

## return last iteration

return iteration

## BARDIAUX 2.2
def _checkSpectrumData(self, spectrum):

k = spectrum.getExperimentData().values()
return len([v for v in k if v <> 0.0]) == len(k)

def run_iteration(self, molecule, last_iteration, iteration,
is_first_iteration = 0):
"""
last_iteration is introduced, to provide time-shifted
pickling of restraint lists (i.e. while another
iteration is on its way)
"""

from time import sleep

check_type(molecule, 'Molecule')
check_type(last_iteration, 'Iteration')

## if all iterations shall be stored in memory ...

## self.iterations[iteration.getNumber()] = iteration

## apply settings of target iteration to all sub-modules
self._updateIterationSettings(iteration)

## get structure ensemble calculated in last iteration
ensemble = last_iteration.getStructureEnsemble()

## We now analyse the structure ensemble of the last
## iteration to generate a new set of restraints which
## will then be used to calculate a new ensemble of
## structures for the target iteration.

spectra = last_iteration.getPeaks().keys()
spectra.sort()

# BARDIAUX 2.2
# add CCPN DistanceRestraints
constraint_lists = last_iteration.getDistanceRestraints().keys()
constraint_lists.sort()

# Malliavin/Bardiaux rMat
cs = self.calibrator.getSettings()

for s in spectra:
if cs['relaxation_matrix'] <> NO and not self._checkSpectrumData(s):
self.warning("Missing experimental data for spectrum \"%s\". " % s.getName())
self.warning("Spin-diffusion correction will be disabled.\n")

cs['relaxation_matrix'] = NO
continue

if cs['relaxation_matrix'] == YES and ensemble is not None:

from NOEModel import SpinDiffusionCorrection

self.model = SpinDiffusionCorrection()
self.calibrator.setModel(self.model)
self.model.prepare(molecule, ensemble)

# initialize matrix
for spectrum in spectra:
self.model.setIntensityMatrix(spectrum)

spectra += constraint_lists

for spectrum in spectra:

self.message('Calibrating spectrum "%s"...' % spectrum.getName())

## get peaks stored in current iteration
## BARDIAUX 2.2 ConstraintList
if is_type(spectrum, 'ConstraintList'):
peaks = iteration.getDistanceRestraints(spectrum)
else:
peaks = iteration.getPeaks(spectrum)

##
## Calibration
##
## If we do not have a seed structure ensemble, we
## perform a very simple calibration
##

## set spectrum-specific parametes for Calibrator


self._updateSpectrumSettings(spectrum)

if ensemble is None:
factor = self.doCalibration(peaks, None)
new_non_violated = None

else:

## Calculate initial calibraton factor.

factor = self.doCalibration(peaks, ensemble)

## Calculate upper/lower bounds for restraints of current
## iteration.

t = clock()

self.calculateBounds(factor, peaks, ensemble = ensemble)

s = '1st calibration and calculation of new ' + \
'distance-bounds done (calibration factor: %e)'
self.message(s % factor, verbose_level = VL_LOW)
self.debug('Time: %ss' % str(clock() - t))

##
## Violation Analysis
##
## Assess every restraint regarding its degree of violation.
##
## Violated restraints will be disabled for the
## current iteration and thus will not be used during
## structure calculation.
##

t = clock()

violated, non_violated, new_non_violated = \
self.doViolationAnalysis(peaks, ensemble)

## Augment set of non-violated restraints

if new_non_violated:
non_violated += new_non_violated

n = len(peaks)
n_viol = len(violated)
p_viol = n_viol * 100. / n

s = 'Violation analysis done: %d / %d restraints ' + \
'(%.1f %%) violated.'

self.message(s % (n_viol, n, p_viol))

if new_non_violated:
s = 'Number of valid restraints has been increased ' + \
'by %d (%.1f%%) after applying a bound-correction.'

p_new = len(new_non_violated) * 100. / n

self.message(s % (len(new_non_violated), p_new))

self.debug('Time: %ss' % str(clock()-t))

##
## 2nd calibration - wrt to non-violated restraints.
## If no restraints have been violated, we use the
## 1st calibration factor.
## Again, we do not store results.
##

if non_violated:
factor = self.doCalibration(non_violated, ensemble)

##
## Activate restraints explicitly.
## We consider a restraint as active, if it has
## not been violated or if its reference cross-peak is
## 'reliable'.
##

for r in peaks:
if not r.analysis.isViolated() or \
r.getReferencePeak().isReliable():
r.isActive(1)
else:
r.isActive(0)

## Store final calibration factor for current iteration.
iteration.setCalibrationFactor(spectrum, factor)

## Calculate upper/lower bounds for restraint-list
## used in the current iteration. I.e. these bounds
## will be used to calculated the structures.

t = clock()

self.calculateBounds(factor, peaks, new_non_violated, ensemble = ensemble)

s = 'Final calibration and calculation of new distance-bounds' + \
' done (calibration factor: %e).' % factor
self.message(s, verbose_level = VL_LOW)
self.debug('Time: %ss' % str(clock() - t))

##
## Partial assignment for restraint-list used in
## current iteration, i.e. for all restraints:
## Calculate weight for every contribution
## and (depends possibly on partial analyser
## settings) throw away 'unlikely' contributions.
##
## If we do not have an ensemble, all contributions
## are activated.
##

t = clock()

# BARDIAUX 2.2
# if we are using the NA, we do not filter the contributions yet
# if the spectrum is a ConstraintList, check if needs to be filtered
NA = self.network_anchoring
ns = NA.getSettings()
do_filter = ns['enabled'] == NO

if is_type(spectrum, 'ConstraintList'):
do_filter = do_filter or \
spectrum.getListSource()['run_network_anchoring'] == NO

do_filter = do_filter and \
spectrum.getListSource()['filter_contributions'] == YES

self.contribution_assigner.assign(peaks, ensemble, filter_contributions = do_filter)

self.message('Partial assignment done.', verbose_level = VL_LOW)
self.debug('Time: %ss' % str(clock() - t))

# BARDIAUX 2.2 NA
NA = self.network_anchoring
ns = NA.getSettings()

if ns['enabled'] == YES:
NA.run(iteration)

# filter contributions
for spectrum in spectra:

# CCPN ConstraintList may not be filtered
if is_type(spectrum, 'ConstraintList'):
if spectrum.getListSource()['filter_contributions'] == NO:
continue
peaks = iteration.getDistanceRestraints(spectrum)
else:
peaks = iteration.getPeaks(spectrum)

self.contribution_assigner.filter_contributions(peaks)


##
## Merge spectra
##

self.mergePeakLists(iteration)


## to known when the structure generation has been completed.
self.__done = 0

## Display some cache statistics
if AriaBaseClass.cache and ensemble is not None:
cache = ensemble._StructureEnsemble__cache
hit_rate = float(cache['hit']) * 100. / cache['total']
self.debug('StructureEnsemble cache hit-rate: %.1f%%' % hit_rate)

##
## Run structure calculation
##

self.startStructureCalculation(iteration, molecule)

##
## Perform analysis wrt to last iteration
## Dump report files
##

if not is_first_iteration:

## apply settings of last iteration to sub-modules
self._updateIterationSettings(last_iteration)

self.writeReports(last_iteration)

## Restore old settings
self._updateIterationSettings(iteration)

## Wait for completion of structure generation.
self.message('Waiting for completion of structure calculation...')

try:

while not self.__done:
sleep(1.)

except KeyboardInterrupt:
self.__done = KEYBOARD_INTERRUPT

if self.__done in (MISSING_STRUCTURES, KEYBOARD_INTERRUPT):

## shut down job-manager

from Singleton import ProjectSingleton

project = ProjectSingleton()
engine = project.getStructureEngine()
engine.getJobScheduler().shutdown()

if self.__done == MISSING_STRUCTURES:

## BARDIAUX 2.2
# get list of missing structures
missing = engine.missingStructures()
msg = 'Structure calculation failed for structure %d.\n' #Please check your setup and the CNS output files for errors.'
err_msg = ''
for m in missing:
err_msg += msg % m

err_msg += 'Please check your setup and the CNS output files for errors.'

## err_msg = 'Structure calculation failed. ' + \
## 'Some structures are missing. ' + \
## 'Please check your setup and/or CNS output files.'

else:
err_msg = 'Interrupted by user.'

self.error(StandardError, err_msg)

## Return current, most recently calculated iteration
return iteration

def run_protocol(self, molecule, iteration):

check_type(molecule, 'Molecule')
check_type(iteration, 'Iteration')

from copy import copy
from Singleton import ProjectSingleton

infra = self.getInfrastructure()
project = ProjectSingleton()
engine = project.getStructureEngine()

msg = '---------------------- Iteration %d -----------------------\n'

is_first_it = 1

n_iterations = len(self.getSettings()['iteration_settings'])

t_main_protocol = clock()

n_first = iteration.getNumber()

for it_number in range(n_first, n_iterations - 1):

## create new iteration

t = clock()

target = copy(iteration)
target._setNumber(iteration.getNumber() + 1)

self.debug('New iteration created: %ss' % str(clock() - t))

self.message(msg % (target.getNumber()))

## Run calculation

t = clock()

iteration = self.run_iteration(molecule, iteration, target,
is_first_it)

## cleanup

if it_number < n_iterations - 1:
it_path = infra.get_iteration_path(it_number)
engine.cleanup(it_path)

is_first_it = 0

self.message('Iteration %s done.' % target.getNumber())

self.debug('Time: %ss' % str(clock() - t))

self.debug('Total time for all iterations: %ss' % \
str(clock() - t_main_protocol))

## Write reports for final iteration
self.writeReports(iteration)
self.writeFinalReports(iteration)

## Return last iteration
return iteration

def writeFinalReports(self, iteration):

check_type(iteration, 'Iteration')

import os, MolMol
from Singleton import ProjectSingleton

## write MOLMOL file-list "file.nam"

molmol_path = self._writePDBFileList(iteration)

if molmol_path is None:
msg = 'Could not write MOLMOL file-list.'
self.warning(msg)
return

## if that worked, attempt to write restraint-lists

project = ProjectSingleton()
s = project.getReporter()['molmol']

if s['enabled'] in (YES, GZIP):

## restraint files

restraints = iteration.getPeakList()
not_merged = [r for r in restraints if not r.isMerged()]

gzip = {YES: 0, GZIP: 1}[s['enabled']]

MolMol.write_noe_restraints(not_merged, molmol_path,
gzip = gzip)
self.message('MOLMOL lower and upper bound (.lol, .upl) ' + \
'files written.')

class ProtocolXMLPickler(XMLBasePickler):

order = ['floating_assignment', 'iteration',
'water_refinement']

def _xml_state(self, x):

e = XMLElement(tag_order = self.order)

s = x.getSettings()

e.iteration = tuple(s['iteration_settings'].values())
e.water_refinement = s['water_refinement']
e.floating_assignment = s['floating_assignment']

return e

def load_from_element(self, e):

from tools import as_tuple

s = ProtocolSettings()

it_settings = as_tuple(e.iteration)

[s.addIterationSettings(i) for i in it_settings]
s['water_refinement'] = e.water_refinement
s['floating_assignment'] = str(e.floating_assignment)

d = s['iteration_settings']

it_numbers = d.keys()

if min(it_numbers) <> 0:

self.warning('Iteration numbering does not start at 0, renumbering ...')

_min = min(it_numbers)

for number, it_settings in d.items():

new_number = number - _min

it_settings['number'] = new_number
d[new_number] = it_settings

del d[number]

p = Protocol(s)

return p

IterationSettings._xml_state = IterationSettingsXMLPickler()._xml_state
Protocol._xml_state = ProtocolXMLPickler()._xml_state

Top of Message | Previous Page | Permalink

JiscMail Tools


RSS Feeds and Sharing


Advanced Options


Archives

May 2024
April 2024
March 2024
February 2024
January 2024
December 2023
November 2023
October 2023
September 2023
August 2023
July 2023
June 2023
May 2023
April 2023
March 2023
February 2023
January 2023
December 2022
November 2022
October 2022
September 2022
August 2022
July 2022
June 2022
May 2022
April 2022
March 2022
February 2022
January 2022
December 2021
November 2021
October 2021
September 2021
August 2021
July 2021
June 2021
May 2021
April 2021
March 2021
February 2021
January 2021
December 2020
November 2020
October 2020
September 2020
August 2020
July 2020
June 2020
May 2020
April 2020
March 2020
February 2020
January 2020
December 2019
November 2019
October 2019
September 2019
August 2019
July 2019
June 2019
May 2019
April 2019
March 2019
February 2019
January 2019
December 2018
November 2018
October 2018
September 2018
August 2018
July 2018
June 2018
May 2018
April 2018
March 2018
February 2018
January 2018
December 2017
November 2017
October 2017
September 2017
August 2017
July 2017
June 2017
May 2017
April 2017
March 2017
February 2017
January 2017
December 2016
November 2016
October 2016
September 2016
August 2016
July 2016
June 2016
May 2016
April 2016
March 2016
February 2016
January 2016
December 2015
November 2015
October 2015
September 2015
August 2015
July 2015
June 2015
May 2015
April 2015
March 2015
February 2015
January 2015
December 2014
November 2014
October 2014
September 2014
August 2014
July 2014
June 2014
May 2014
April 2014
March 2014
February 2014
January 2014
December 2013
November 2013
October 2013
September 2013
August 2013
July 2013
June 2013
May 2013
April 2013
March 2013
February 2013
January 2013
December 2012
November 2012
October 2012
September 2012
August 2012
July 2012
June 2012
May 2012
April 2012
March 2012
February 2012
January 2012
December 2011
November 2011
October 2011
September 2011
August 2011
July 2011
June 2011
May 2011
April 2011
March 2011
February 2011
January 2011
December 2010
November 2010
October 2010
September 2010
August 2010
July 2010
June 2010
May 2010
April 2010
March 2010
February 2010
January 2010
December 2009
November 2009
October 2009
September 2009
August 2009
July 2009
June 2009
May 2009
April 2009
March 2009
February 2009
January 2009
December 2008
November 2008
October 2008
September 2008
August 2008
July 2008
June 2008
May 2008
April 2008
March 2008
February 2008
January 2008
December 2007
November 2007
October 2007
September 2007
August 2007
July 2007
June 2007
May 2007
April 2007
March 2007
February 2007
January 2007
December 2006
November 2006
October 2006
September 2006
August 2006
July 2006
June 2006
May 2006
April 2006
March 2006
February 2006
January 2006
December 2005
November 2005
October 2005
September 2005
August 2005
July 2005
June 2005
May 2005
April 2005
March 2005
February 2005
January 2005
December 2004
November 2004
October 2004
September 2004
August 2004
July 2004
June 2004
May 2004
April 2004
March 2004
February 2004
January 2004
December 2003
November 2003
October 2003
September 2003


JiscMail is a Jisc service.

View our service policies at https://www.jiscmail.ac.uk/policyandsecurity/ and Jisc's privacy policy at https://www.jisc.ac.uk/website/privacy-notice

For help and support help@jisc.ac.uk

Secured by F-Secure Anti-Virus CataList Email List Search Powered by the LISTSERV Email List Manager