I have a problem related to measuring likelihood between -an observed presence absence dataset (containing 0 or 1) -a predicted simulation matrix of the same dimensions (containing values from 0 to 1)
This must be a common problem but I am struggling to find the answer in the literature. Within the simulation model I have a parameter 'm' which I can alter to find the best fit (maximum likelihood). Currently I just use a 'sum of squares of difference' to measure likelihood. ie likelihood = sum (obs-pred)^2 This is then very easy to find (using numerical optimisation techniques) the value of 'm' which gives my maximum likelihood (least sum of squares) However I do not think my likelihood function is the correct one to be using for this purpose. Firstly, if sum of squares is the correct method, maybe I should be taking the square root of the likelihood (makes no difference) and possibly the 'mean' values of the datasets may need to be included in my calculatons. However, sum of squares suggests my data are normally distributed (which it is clearly not) Obs (boolean O or 1) Pred (beta O to 1) Difference (beta -1 to 1) My guess is that I should be using a beta (or uniform) defined likelihood measure. Or maybe just a simple transformation. Any help greatly appreciated Mark _________________________________________________________________ [[alternative HTML version deleted]] ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.