On Nov 17, 2010, at 7:33 AM, Martin Tomko wrote: > Dear all, > I am having a hard time to figure out a suitable test for the match between > two nominal classifications of the same set of data. > I have used hierarchical clustering with multiple methods (ward, k-means,...) > to classify my dat into a set number of classesa, and I would like to compare > the resulting automated classification with the actual - objective benchmark > one. > So in principle I have a data frame with n columns of nominal > classifications, and I want to do a mutual comparison and test for > significance in difference in classification between pairs of columns. > > I just need to identify a suitable test, but I fail. I am currently exploring > the possibility of using Cohen's Kappa, but I am open to other suggestions. > Especially the fact that kappa seems to be moslty used on failible, human > annotators seems to bring in limitations taht do not apply to my automatic > classification. > Any help will be appreciated, especially if also followed by a pointer to an > R package that implements it. > > Thanks > Martin
In addition to Matt's comments, you might want to consider marginal homogeneity tests. There are extensions of the pairwise McNemar test to greater than two categories. Some online information is here: http://www.john-uebersax.com/stat/mcnemar.htm and there is the ?mh_test implemented in the 'coin' package on CRAN. HTH, Marc Schwartz ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.