Hi - what Mark says makes a lot of sense. I would just add a couple of minor points: - You should definitely make sure you're using FSL 4.1 as it has more liberal masking than 4.0 (for these very reasons). - Not sure what Vince is thinking as the spatial smoothing is effectively done at the same stage in FSL and SPM, AFAIK. - But yes, reducing the masking threshold will definitely make the masking more liberal. Doing this is fine. Cheers. On 14 May 2009, at 08:49, Mark Jenkinson wrote: > Hi Stefan, > > The problem has no easy answer I'm afraid. > It isn't something that we've really come across very much, although > we have been thinking about ways of better incorporating this in > future releases now. > > There are three potential things that you could do. > > One is to do what David suggests and just make the masks bigger (e.g. > by lowering the %brain/background threshold) but this may include > voxels that do not contain useful signal and hence bias your results, > although with such a high N, this might not be too bad. > > The second thing to do is to include some voxel-specific regressors > which model out the effects of the missing data points. You would > also need to increase the masks to get these voxels included, and > you would need one regressor per number of subjects that you wanted > to reinstate. For example, to include all voxels where one subject > only was missing you would need one voxel-wise regressor which > contained zeros for all timepoints except a single one where there > was missing data. This doesn't have to be the same subject in each > voxel - this is the advantage of the voxel-wise regressors - so that > you would have certain voxels containing an all zero regressor (if > they had data for all timepoints) and others with a single one in the > appropriate slot corresponding to the subject which was missing > *for that voxel*. Note that timepoints = subject index in the above. > This is a nicer mathematical solution but is practically a little > difficult > to do and if you wanted to include voxels where two subjects were > missing then you'd need another voxel-wise regressor and so on. > Therefore if you wanted to include voxels where you had N=300 > or 299 or 298 or ... 270, then you'd need to make 30 voxel-wise > regressors and include them all which is cumbersome and may > slow the analysis down a fair bit. > > The third solution would be to use the outlier detection mechanism > to do the work for you. This again requires enlarging the masks for > each subject (to include all the voxels you want to - and it may be > better to do this by replacing the mask files in reg_standard directly > with a common mask that you want). Once you've done this you > would need to replace values in the cope images (again in > reg_standard) > which were previously outside the mask (and hence would currently > be zero) with values which are clearly outliers. You should be able > to work out a sensible outlier value by looking at your current > analyses and seeing how much the valid copes vary. If you do this > the outlier detection should hopefully identify these subjects as > outliers in those voxels and hence effectively remove them from > the analysis. I must say that I haven't tried this solution (or in > fact > the previous one) but in principle it should work. > > Sorry I haven't got a nicer and easier solution for you. > > All the best, > Mark > > > On 14 May 2009, at 04:10, David V. Smith wrote: > >> Hi Stefan, >> >> Which version of FSL are you using? I believe 4.0 was a bit too >> conservative when generating the masks (see previous posts on this >> matter), but the newer versions don’t really have a problem with >> this. >> >> But, regardless of what version you’re using, once a voxel is gone, >> it’s gone for everyone. So, even if you have just one bad mask in >> your dataset, it will mess up the others. >> >> I think your approach to overcoming this (i.e., lowering the %brain/ >> background threshold) is fine. Just make sure you do it for everyone. >> >> Hope this helps, >> David >> >> >> -- >> David V. Smith >> Graduate Student, Huettel Lab >> Center for Cognitive Neuroscience >> Duke University >> Box 90999 >> Durham, NC 27708 >> Lab phone: (919) 668-3635 >> Lab website: http://www.duke.edu/web/mind/level2/faculty/huettel/ >> >> From: FSL - FMRIB's Software Library [mailto:[log in to unmask]] On >> Behalf Of Stefan Ehrlich >> Sent: Wednesday, May 13, 2009 7:52 PM >> To: [log in to unmask] >> Subject: [FSL] problems due to concatenated masks in GFEAT of a >> very large study - 3rd posting - no answer? >> >> Dear FSL'ers, >> >> >> >> I had posted that earlier. This is the 3rd posting. It would be >> *very* helpful if someone could give me some advice on that. Please >> also let me also know If the problem is not described clearly >> enough or if there is no easy answer. >> >> >> >> I am running FEAT on a large fMRI dataset (n > 300). Each subject >> has 3 runs consisting of 16 blocks of a memory paradigm. After >> hundreds of hours of computing I have processed all first- and >> second level-analyses (from here on referred to as cross-runs). I >> have checked the registrations and individual as well as cross-run >> activation-maps (retrieval versus fixation) for a subset of >> subjects and all seems fine. As a next step I ran a “Single-Group >> Average” over all 300 cross-runs just to get an impression of the >> overall activation patterns. The results were in line with previous >> studies but in the top slices of the brain (horizontal slices) as >> well as in a rim covering the parts of the brain closest to the >> skull there was no activation whatsoever. I found out that this was >> probably due to the mask.nii of the gfeat. This group mask did not >> include the top slices and the outer rim . >> >> Subsequently I went back and checked all cross-run mask.nii and >> identified a very few masks which were missing several top slices >> (due to bad positioning in the scanner, I guess). After deleting >> these subjects from the overall analysis my final results and the >> “Single-Group Average” mask looks much better. However, the outer >> rim is still missing. >> >> It seems like FEAT concatenates all cross-run masks and does not >> include voxels which have a missing value in any single mask. Vince >> Calhoun told me on the phone that this might be due to the fact >> that FSL smoothes relatively late in the processing stream (in >> contrast to SPM). Concequently I went back and changed the brain/ >> background threshold (brain_thresh) from 10 to 1 and rerun a few >> subjects which had slightly impaired cross-run masks. With the new >> threshold more voxels get included. Do you think that is an >> approbriate approach? Has anybody experienced that problem before? >> Are there other solutions? >> >> >> >> Thank you so much for your thoughts! >> >> >> >> Stefan >> >> >> >> > --------------------------------------------------------------------------- Stephen M. Smith, Professor of Biomedical Engineering Associate Director, Oxford University FMRIB Centre FMRIB, JR Hospital, Headington, Oxford OX3 9DU, UK +44 (0) 1865 222726 (fax 222717) [log in to unmask] http://www.fmrib.ox.ac.uk/~steve ---------------------------------------------------------------------------