Fiber density direct comparison across groups

Hello

I was trying to compare the fiber density from subjects across two groups. However, I started to doubt whether this is a valid way to do so…

I followed the HCP connectome tutorital to process my subjects, both controls and patients, with SIFT to filter the tracts. In the end all my subjects have 1million tracts over the whole brain. I calculated fiber density through dividing the fiber counts between ROIs by the volume of the ROIs and try to comprare them across groups.

My question is, 1, Most of the publications calculating fiber density use it to generate graph theory based measures. Can I do simple comparions between them instead of these graph measures? (I would say yes since SIFT claimed to help improve the biological accuracy of structural connectivity.)

  1. If yes for the 1st question, then is the fiber density i calculated is the correct measure? Although it is normalized with the ROI volume, the way the whole-brain connectome was generated may mess with it. After SIFT, every subject has the same number of whole-brain tracts, this may cause bias due to the volume of the brain. E.G. a subject with 1000 mm3 brain size should not have the same number of tracts with a subject with 10000 mm3. Does it make sense? Does it mean that if I want to compare the fiber density, I need to multiply the whole brain volume?

Any comments are appreciated.

Thanks!

Jidan

Hi,

Sorry for taking longer than usual to get to this one. I’m writing a new page for the documentation discussing some of these topics, while simultaneously trying to avoid sounding like a broken record regarding a paper I should have published 2 years ago. There will likely be a blog post announcing it once it appears (hopefully next week).

Cheers
Rob

Hi Jidan,

Hopefully you caught my blog post a couple of days ago that discussed some of the points you raised. But I’ll address some of them more directly here:

I calculated fiber density through dividing the fiber counts between ROIs by the volume of the ROIs …

This in particular was one of the more focused points of my blog post; if you’re looking for a more meaningful measure of fibre density, this step alone is already diluting that measurement / interpretation.

Most of the publications calculating fiber density use it to generate graph theory based measures. Can I do simple comparions between them instead of these graph measures?

Absolutely. This is very much an intended application of SIFT and related methods: a direct comparison of fibre density of particular pathways. The disadvantage of such an approach is the need for an a priori hypothesis: you can’t really test every edge in the connectome independently for an effect, since the correction for multiple comparisons will destroy your statistical power. You should also note that the inter-subject variance in the quantification of fibre density in individual edges is quite high; see for example the scatterplots I generated in this paper.

Once it’s formally published we will add the NBS2 method to MRtrix3, which provides somewhat of a compromise between individual connection testing and network analysis.

If yes for the 1st question, then is the fiber density i calculated is the correct measure? Although it is normalized with the ROI volume, the way the whole-brain connectome was generated may mess with it. After SIFT, every subject has the same number of whole-brain tracts, this may cause bias due to the volume of the brain. E.G. a subject with 1000 mm3 brain size should not have the same number of tracts with a subject with 10000 mm3. Does it make sense? Does it mean that if I want to compare the fiber density, I need to multiply the whole brain volume?

Yes, this is a topic that crops up repeatedly, and I should have published on a long time ago. I tend to refer to this as connection density normalisation; it’s a bit like inter-subject image intensity normalisation, except that you’re trying to make your measure of connection density comparable between subjects.

This issue requires a little more finesse than simply identifying factors that may influence the quantification of connection density, and multiplying / dividing as seems appropriate. For instance: if you had gross differences in brain volumes, this would already influence the parcellation node volumes, and therefore your initial fibre density estimates if you divide the streamline count in each edge by the node volumes - subsequently scaling these estimates by brain volume could either partially cancel out the intended effect, or amplify the effect such that it itself becomes a systematic bias, rather than compensating for a bias.

Some of the arguments I made in the blog post regarding scaling edge densities by node volumes - particularly, the units of the quantification and hence interpretability of the measure, and the formulation of a hypothesis - are also relevant for mechanisms such as scaling for whole-brain volume. So I would suggest re-reading those in this context.

In general I tend to recommend that people stick with simply fixing the number of streamlines per subject - at least until I publish an alternative :stuck_out_tongue: . It’s already common in the field, it shouldn’t be too hard to get past reviewers, and its limitations can be acknowledged appropriately.

Thanks for your patience!
Rob

Hi Robert

Thanks for your reply! I do have read your blog post. There 's another related paper published by Yeh from your group recently (the one accepted by Neuroimage). By reading that, and with your current reply it helped a lot on the understanding of the streamlines.

I agree that streamline counts between ROI pairs make more sense because every subject has the same number of total streamlines. In this case the streamline count between ROI pairs is “streamline density” per fixed whole-brain total number of streamlines (referred as your connection density normalization). And yes, direct comparison of the streamline counts does suffer from multiple comparison correction due to the massive pairs of ROIs across the whole brain. I do have a priori hypothesis for this analysis so that my results would not be washed by the multiple comparison correction.

Thanks again!
Jidan