On 07/01/2010 03:29 AM, Dr. David Kirkby wrote:
On 03/ 1/10 12:23 AM, Sharpie wrote:


John Maindonald wrote:

I came across this notice of an upcoming webinar. The issues identified
in the
first paragraph below seem to me exactly those that the R project is
designed
to address. The claim that "most research software is barely fit for
purpose
compared to equivalent systems in the commercial world" seems to me not
quite accurate! Comments!

It can be argued that this is a reporting bias. Whenever I inform people doing epidemiology with Excel about Ian Buchan's paper on Excel errors:

http://www.nwpho.org.uk/sadb/Poisson%20CI%20in%20spreadsheets.pdf

there is a sort of reflexive disbelief, as though something as widely used as Excel could not possibly be wrong. That is to say, most people using commercial software, especially the sort that allows them to follow a cookbook method and get an acceptable (to supervisors, journal editors and paymasters) result simply accept it without question.

The counterweight to the carefree programming style employed by many researchers (I include myself) is the multitude of enquiring eyes that find our mistakes, and foster a continual refinement of our programs. I just received one this evening, about yet another thing that I had never considered, perfect agreement by rating methods in a large trial. Thus humanity bootstraps upward. My AUD0.02

Jim

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to