Hi all,

Thanks for the replies (including off list).  I have since resolved the
discrepant results. I believe it has to do with R's scoping rules - I had an
object called 'labs' and a variable in the dataset (DATA) called 'labs', and
apparently (to my surprise), when I called this:

lmer(Y~X + (1|labs),dataset=DATA)

lmer was using the object 'labs' rather than the object 'DATA$labs'. Is this
expected behavior??

This would have been fine, except I had reordered DATA in the meantime!

Best,

JJ

On Tue, Aug 17, 2010 at 7:17 PM, Mitchell Maltenfort <mmal...@gmail.com>wrote:

> One difference is that the random effect in lmer is assumed --
> implicitly constrained, as I understand it -- to
> be a bell curve.  The fixed effect model does not have that constraint.
>
> How are the values of "labs" effects distributed in your lm model?
>
> On Tue, Aug 17, 2010 at 8:50 PM, Johan Jackson
> <johan.h.jack...@gmail.com> wrote:
> > Hello,
> >
> > Setup: I have data with ~10K observations. Observations come from 16
> > different laboratories (labs). I am interested in how a continuous
> factor,
> > X, affects my dependent variable, Y, but there are big differences in the
> > variance and mean across labs.
> >
> > I run this model, which controls for mean but not variance differences
> > between the labs:
> > lm(Y ~ X + as.factor(labs)).
> > The effect of X is highly significant (p < .00001)
> >
> > I then run this model using lme4:
> > lmer(Y~ X + (1|labs)) #controls for mean diffs bw labs
> > lmer(Y~X + (X|labs)) #and possible slope heterogeneity bw labs.
> >
> > For both of these latter models, the effect of X is non-significant (|t|
> <
> > 1.5).
> >
> > What might this be telling me about my data? I guess the second (X|labs)
> may
> > tell me that there are big differences in the slope across labs, and that
> > the slope isn't significant against the backdrop of 16 slopes that differ
> > quite a bit between each other. Is that right? (Still, the enormous drop
> in
> > p-value is surprising!). I'm not clear on why the first (1|labs),
> however,
> > is so discrepant from just controlling for the mean effects of labs.
> >
> > Any help in interpreting these data would be appreciated. When I first
> saw
> > the data, I jumped for joy, but now I'm muddled and uncertain if I'm
> > overlooking something. Is there still room for optimism (with respect to
> X
> > affecting Y)?
> >
> > JJ
> >
> >        [[alternative HTML version deleted]]
> >
> > ______________________________________________
> > R-help@r-project.org mailing list
> > https://stat.ethz.ch/mailman/listinfo/r-help
> > PLEASE do read the posting guide
> http://www.R-project.org/posting-guide.html
> > and provide commented, minimal, self-contained, reproducible code.
> >
>

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to