Hello,

Try ?aggregate, it's probably faster.
With a made up data.frame, since you haven't provided us with a dataset,

simout.s1 <- data.frame(SID = rep(LETTERS[1:2], 10),
                DOSENO = rep(letters[1:4], each = 5),
                value = rnorm(20))

res2 <- aggregate(simout.s1$value, list(simout.s1$SID, simout.s1$DOSENO), function(x)x[NROW(x)])
names(res2) <- names(simout.s1)


Use dput to post a data example. Something like the following.

dput(head(simout.s1, 50))  #paste the output of this in your next mail


Hope this helps,

Rui Barradas



Citando Jun Shen <jun.shen...@gmail.com>:

Dear list,

I have the following line of code to extract the last line of the split
data and put them back together.

do.call(rbind,lapply(split(simout.s1,simout.s1[c('SID','DOSENO')]),function(x)x[nrow(x),]))

the problem is when  have a huge dataset, it takes too long to run.
(actually it's > 3 hours and it's still running).

The dataset is pretty big. I have 200,000 unique SID and 4 DOSENO, so
totally 800,000 split dataset. Is there anyway to speed it up? Thanks.

Jun

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

______________________________________________
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to