Hi Luc,

I will look at this in dfp. I saw the package and thought it was
davidon-fletcher-powel. ;-)

I attempted to do what you suggest with BigDecimal. Everything was okay and
there were some marginal benefits with doing so, but I thought the hassle
was not worth it (at least with BigDecimal).

I will try with dfp!

Thank you,

-Greg

On Fri, Jul 15, 2011 at 1:56 AM, Luc Maisonobe <luc.maison...@free.fr>wrote:

> Le 15/07/2011 02:37, Greg Sterijevski a écrit :
>
>  The usual issues with numerical techniques, how you calculate (c * x + d *
>> y)/e matters...
>> It turns out that religiously following the article and defining c_bar  =
>> c
>> / e is not a good idea.
>>
>> The Filippelli data is still a bit dicey. I would like to resolve where
>> the
>> error is accumulating there as well. That's really the last thing
>> preventing
>> me from sending the patch with the Miller-Gentlemen Regression to Phil.
>>
>
> I don't know whether this is feasible in your case, but when trying to find
> this kind of numerical errors, I found useful to just redo the computation
> in parallel to high precision. Up to a few months ago, I was simply doing
> this using emacs (yes, emacs rocks) configured with 50 significant digits?
> Now it is easier since we have our own dfp package in [math].
>
> Luc
>
>
>
>> -Greg
>>
>> On Thu, Jul 14, 2011 at 1:18 PM, Ted Dunning<ted.dunn...@gmail.com>
>>  wrote:
>>
>>  What was the problem?
>>>
>>> On Wed, Jul 13, 2011 at 8:33 PM, Greg Sterijevski<gsterijevski@**
>>> gmail.com <gsterijev...@gmail.com>
>>>
>>>> wrote:
>>>>
>>>
>>>  Phil,
>>>>
>>>> Got it! I fit longley to all printed values. I have not broken
>>>>
>>> anything...
>>>
>>>> I
>>>> need to type up a few loose ends, then I will send a patch.
>>>>
>>>> -Greg
>>>>
>>>> On Tue, Jul 12, 2011 at 2:35 PM, Phil Steitz<phil.ste...@gmail.com>
>>>> wrote:
>>>>
>>>>  On 7/12/11 12:12 PM, Greg Sterijevski wrote:
>>>>>
>>>>>> All,
>>>>>>
>>>>>> So I included the wampler data in the test suite. The interesting
>>>>>>
>>>>> thing,
>>>>
>>>>> is
>>>>>
>>>>>> to get clean runs I need wider tolerances with OLSMultipleRegression
>>>>>>
>>>>> than
>>>>
>>>>> with the version of the Miller algorithm I am coding up.
>>>>>>
>>>>> This is good for your Miller impl, not so good for
>>>>> OLSMultipleRegression.
>>>>>
>>>>>> Perhaps we should come to a consensus of what good enough is? How
>>>>>>
>>>>> close
>>>
>>>> do
>>>>>
>>>>>> we want to be? Should we require passing on all of NIST's 'hard'
>>>>>>
>>>>> problems?
>>>>>
>>>>>> (for all regression techniques that get cooked up)
>>>>>>
>>>>>>  The goal should be to match all of the displayed digits in the
>>>>> reference data.  When we can't do that, we should try to understand
>>>>> why and aim to, if possible, improve the impls.   As we improve the
>>>>> code, the tolerances in the tests can be improved.  Characterization
>>>>> of the types of models where the different implementations do well /
>>>>> poorly is another thing we should aim for (and include in the
>>>>> javadoc).  As with all reference validation tests, we need to keep
>>>>> in mind that a) the "hard" examples are designed to be numerically
>>>>> unstable and b) conversely, a handful of examples does not really
>>>>> demonstrate correctness.
>>>>>
>>>>> Phil
>>>>>
>>>>>> -Greg
>>>>>>
>>>>>>
>>>>>
>>>>> ------------------------------**------------------------------**
>>>>> ---------
>>>>> To unsubscribe, e-mail: 
>>>>> dev-unsubscribe@commons.**apache.org<dev-unsubscr...@commons.apache.org>
>>>>> For additional commands, e-mail: dev-h...@commons.apache.org
>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>
> ------------------------------**------------------------------**---------
> To unsubscribe, e-mail: 
> dev-unsubscribe@commons.**apache.org<dev-unsubscr...@commons.apache.org>
> For additional commands, e-mail: dev-h...@commons.apache.org
>
>

Reply via email to