Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix typo at distances.rst #417

Merged
merged 2 commits into from
Aug 27, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/source/distances.rst
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ Estimating dissimilarities
==========================
The second step in RSA is to calculate a dissimilarity of distance measure from the data.

To calculate all pairwise distance, one needs to define the appropriate dataset (see :ref:`datasets`) and use ``rdm.calc.calc_rdm`` to caluclate the appropriate rdm.
To calculate all pairwise distance, one needs to define the appropriate dataset (see :ref:`datasets`) and use ``rdm.calc.calc_rdm`` to calculate the appropriate rdm.

.. sourcecode:: python

Expand Down Expand Up @@ -108,7 +108,7 @@ The really big advantage of this dissimilarity measure is that it is unbiased. I

* You can perform a t-test of the crossnobis estimates against zero, exactly like you would test the classification accuracy of a decoder against chance performance. Thus, you do not need to perform a decoding analysis to determine where there is reliable information, and then conduct an RSA analysis to make inferences about the shape of the representation. By using the crossnobis dissimilarity, you can do so in one step (see also \ref{unbiasedDistanceCorrelation}.

* Unequal noise across conditions can severely bias RDMs computed with normal distances. For example when you have less trials for one condition than another, the pattern for that condition :math:`\bar{\mathbf{b}}_i` has higher noise variance. Therefore the distance to other conditions will be higher, even though the condition only differs by noise. When two conditions are estimated with correlated noise (for example when they are acquired in close temrpora; proximity with fMRI), their distance will be smaller than when they are collected with independent noise. For example, one cannot compare 'normal' distances within an imaging run to distances across imaging runs (due to correlated noise, the former will be usually smaller). Cross-validation removes these biases, making inference much more robust.
* Unequal noise across conditions can severely bias RDMs computed with normal distances. For example when you have less trials for one condition than another, the pattern for that condition :math:`\bar{\mathbf{b}}_i` has higher noise variance. Therefore the distance to other conditions will be higher, even though the condition only differs by noise. When two conditions are estimated with correlated noise (for example when they are acquired in close temporal proximity with fMRI), their distance will be smaller than when they are collected with independent noise. For example, one cannot compare 'normal' distances within an imaging run to distances across imaging runs (due to correlated noise, the former will be usually smaller). Cross-validation removes these biases, making inference much more robust.

* Having a meaningful zero point (i.e. the true patterns are not different) can help in model comparison, as it provides another informative point (Diedrichsen et al., 2020). To exploit this, it is recommended to use the cosine similarity instead of the Pearson correlation for RDM comparison.

Expand Down
Loading