KAPPA estimates Cohen's kappa coefficient and related statistics [...] = kappa(d1,d2); NaN's are handled as missing values and are ignored [...] = kappa(d1,d2,'notIgnoreNAN'); NaN's are handled as just another Label. [kap,sd,H,z,ACC,sACC,MI] = kappa(...); X = kappa(...); d1 data of scorer 1 d2 data of scorer 2 kap Cohen's kappa coefficient point se standard error of the kappa estimate H Concordance matrix, i.e. confusion matrix z z-score ACC overall agreement (accuracy) sACC specific accuracy MI Mutual information or transfer information (in [bits]) X is a struct containing all the fields above For two classes, a number of additional summary statistics including TPR, FPR, FDR, PPV, NPF, F1, dprime, Matthews Correlation coefficient (MCC) or Phi coefficient (PHI=MCC), Specificity and Sensitivity, Youden index (YI) are provided. Note, the positive category must the larger label (in d and c), otherwise the confusion matrix becomes transposed and the summary statistics are messed up. Reference(s): [1] Cohen, J. (1960). A coefficient of agreement for nominal scales. Educational and Psychological Measurement, 20, 37-46. [2] J Bortz, GA Lienert (1998) Kurzgefasste Statistik f|r die klassische Forschung, Springer Berlin - Heidelberg. Kapitel 6: Uebereinstimmungsmasze fuer subjektive Merkmalsurteile. p. 265-270. [3] http://www.cmis.csiro.au/Fiona.Evans/personal/msc/html/chapter3.html [4] Kraemer, H. C. (1982). Kappa coefficient. In S. Kotz and N. L. Johnson (Eds.), Encyclopedia of Statistical Sciences. New York: John Wiley & Sons. [5] http://ourworld.compuserve.com/homepages/jsuebersax/kappa.htm [6] http://en.wikipedia.org/wiki/Receiver_operating_characteristic
Package: nan