Hi Luc, I will look at this in dfp. I saw the package and thought it was davidon-fletcher-powel. ;-)
I attempted to do what you suggest with BigDecimal. Everything was okay and there were some marginal benefits with doing so, but I thought the hassle was not worth it (at least with BigDecimal). I will try with dfp! Thank you, -Greg On Fri, Jul 15, 2011 at 1:56 AM, Luc Maisonobe <luc.maison...@free.fr>wrote: > Le 15/07/2011 02:37, Greg Sterijevski a écrit : > > The usual issues with numerical techniques, how you calculate (c * x + d * >> y)/e matters... >> It turns out that religiously following the article and defining c_bar = >> c >> / e is not a good idea. >> >> The Filippelli data is still a bit dicey. I would like to resolve where >> the >> error is accumulating there as well. That's really the last thing >> preventing >> me from sending the patch with the Miller-Gentlemen Regression to Phil. >> > > I don't know whether this is feasible in your case, but when trying to find > this kind of numerical errors, I found useful to just redo the computation > in parallel to high precision. Up to a few months ago, I was simply doing > this using emacs (yes, emacs rocks) configured with 50 significant digits? > Now it is easier since we have our own dfp package in [math]. > > Luc > > > >> -Greg >> >> On Thu, Jul 14, 2011 at 1:18 PM, Ted Dunning<ted.dunn...@gmail.com> >> wrote: >> >> What was the problem? >>> >>> On Wed, Jul 13, 2011 at 8:33 PM, Greg Sterijevski<gsterijevski@** >>> gmail.com <gsterijev...@gmail.com> >>> >>>> wrote: >>>> >>> >>> Phil, >>>> >>>> Got it! I fit longley to all printed values. I have not broken >>>> >>> anything... >>> >>>> I >>>> need to type up a few loose ends, then I will send a patch. >>>> >>>> -Greg >>>> >>>> On Tue, Jul 12, 2011 at 2:35 PM, Phil Steitz<phil.ste...@gmail.com> >>>> wrote: >>>> >>>> On 7/12/11 12:12 PM, Greg Sterijevski wrote: >>>>> >>>>>> All, >>>>>> >>>>>> So I included the wampler data in the test suite. The interesting >>>>>> >>>>> thing, >>>> >>>>> is >>>>> >>>>>> to get clean runs I need wider tolerances with OLSMultipleRegression >>>>>> >>>>> than >>>> >>>>> with the version of the Miller algorithm I am coding up. >>>>>> >>>>> This is good for your Miller impl, not so good for >>>>> OLSMultipleRegression. >>>>> >>>>>> Perhaps we should come to a consensus of what good enough is? How >>>>>> >>>>> close >>> >>>> do >>>>> >>>>>> we want to be? Should we require passing on all of NIST's 'hard' >>>>>> >>>>> problems? >>>>> >>>>>> (for all regression techniques that get cooked up) >>>>>> >>>>>> The goal should be to match all of the displayed digits in the >>>>> reference data. When we can't do that, we should try to understand >>>>> why and aim to, if possible, improve the impls. As we improve the >>>>> code, the tolerances in the tests can be improved. Characterization >>>>> of the types of models where the different implementations do well / >>>>> poorly is another thing we should aim for (and include in the >>>>> javadoc). As with all reference validation tests, we need to keep >>>>> in mind that a) the "hard" examples are designed to be numerically >>>>> unstable and b) conversely, a handful of examples does not really >>>>> demonstrate correctness. >>>>> >>>>> Phil >>>>> >>>>>> -Greg >>>>>> >>>>>> >>>>> >>>>> ------------------------------**------------------------------** >>>>> --------- >>>>> To unsubscribe, e-mail: >>>>> dev-unsubscribe@commons.**apache.org<dev-unsubscr...@commons.apache.org> >>>>> For additional commands, e-mail: dev-h...@commons.apache.org >>>>> >>>>> >>>>> >>>> >>> >> > > ------------------------------**------------------------------**--------- > To unsubscribe, e-mail: > dev-unsubscribe@commons.**apache.org<dev-unsubscr...@commons.apache.org> > For additional commands, e-mail: dev-h...@commons.apache.org > >