User:Jadrian Miles/Thesis manifesto: probabilistic worldview: Difference between revisions
New page: My dissertation involves solving three problems: # Automatically clustering tractography curves together so that the resulting clusters are neither too small (low reconstruction error, hig... |
Introduction to chi-square fitting |
||
| Line 5: | Line 5: | ||
In each case, our goal is to balance the tradeoff between reconstruction error and model complexity. In order to do so in a principled fashion, we must define, for each case, what is meant by reconstruction error and model complexity. | In each case, our goal is to balance the tradeoff between reconstruction error and model complexity. In order to do so in a principled fashion, we must define, for each case, what is meant by reconstruction error and model complexity. | ||
===<math>\chi^2</math> Fitting=== | |||
One apparent option is <math>\chi^2</math> fitting. Given: | |||
* <math>N</math> observations <math>y_i</math> at independent variables <math>x_i</math> | |||
* The knowledge that the measurement error on each observation is Gaussian with standard deviation <math>\sigma_i</math> | |||
* A model instance with <math>M</math> parameters <math>a_j</math>, which gives a reconstructed "observation" for a given independent variable as <math>y(x_i|a_0 \dots a_{M-1})</math> | |||
we define the <math>\chi^2</math> statistic as follows: | |||
<math>\chi^2 = \sum_{i=0}^{N-1}\left(\frac{y_i - y(x_i|a_0 \dots a_{M-1})}{\sigma_i}\right)^2</math> | |||
This is essentially a normalized error score. The interesting thing that we can do here, though, is compute the probability of seeing an error as bad or worse than the observed error score under a different set of observations drawn from the same underlying distribution, given the degrees of freedom <math>\nu \equiv N - M</math>. This probability is called <math>Q</math>, and is defined as: | |||
<math>Q = 1 - \int_{\chi^2}^{\infty}F(\psi,\nu)\,\mathrm{d}\psi</math> | |||
where <math>F(\psi,\nu)</math> is the [[w:Chi-squared distribution#Cumulative distribution function|CDF of the <math>\chi^2</math> distribution]]. Here's how to interpret <math>Q</math>: | |||
* Really low (<math>Q < 0.001</math>) means that you can't do a much worse job describing the data than you did with your model instance. This means either that your fit is bad (<math>\chi^2</math> is really high) or that you have too many degrees of freedom. | |||
* Really high (<math>Q \approx 1</math>) means that you can't do much better, even given the number of degrees of freedom. This is a suspicious result and may indicate data-fudging, so it shouldn't come up in an automated system. | |||
This means that maximizing <math>Q</math> over all values of <math>M</math> (from very simple to very complex models) should find the "happy medium". | |||
Revision as of 00:40, 5 March 2011
My dissertation involves solving three problems:
- Automatically clustering tractography curves together so that the resulting clusters are neither too small (low reconstruction error, high model complexity) nor too big (high reconstruction error, low model complexity)
- Automatically adjusting macrostructure elements to match input DWIs so that the elements' surfaces are neither too bumpy (low reconstruction error, high model complexity) nor too smooth (high reconstruction error, low model complexity)
- Automatically adjusting microstructure properties within a given region in space to match input DWIs so that the spatial frequency of the microstructure parameters is neither too high (low reconstruction error, high model complexity) nor too low (high reconstruction error, low model complexity)
In each case, our goal is to balance the tradeoff between reconstruction error and model complexity. In order to do so in a principled fashion, we must define, for each case, what is meant by reconstruction error and model complexity.
Fitting
One apparent option is fitting. Given:
- observations at independent variables
- The knowledge that the measurement error on each observation is Gaussian with standard deviation
- A model instance with parameters , which gives a reconstructed "observation" for a given independent variable as
we define the statistic as follows:
This is essentially a normalized error score. The interesting thing that we can do here, though, is compute the probability of seeing an error as bad or worse than the observed error score under a different set of observations drawn from the same underlying distribution, given the degrees of freedom . This probability is called , and is defined as:
where is the CDF of the distribution. Here's how to interpret :
- Really low () means that you can't do a much worse job describing the data than you did with your model instance. This means either that your fit is bad ( is really high) or that you have too many degrees of freedom.
- Really high () means that you can't do much better, even given the number of degrees of freedom. This is a suspicious result and may indicate data-fudging, so it shouldn't come up in an automated system.
This means that maximizing over all values of (from very simple to very complex models) should find the "happy medium".