| agree | Simple and extended percentage agreement |
| anxiety | Anxiety ratings by different raters |
| bhapkar | Bhapkar coefficient of concordance between raters |
| diagnoses | Psychiatric diagnoses provided by different raters |
| finn | Finn coefficient for oneway and twoway models |
| icc | Intraclass correlation coefficient (ICC) for oneway and twoway models |
| iota | iota coefficient for the interrater agreement of multivariate observations |
| kappa2 | Cohen's Kappa and weighted Kappa for two raters |
| kappam.fleiss | Fleiss' Kappa for m raters |
| kappam.light | Light's Kappa for m raters |
| kendall | Kendall's coefficient of concordance W |
| kripp.alpha | calculate Krippendorff's alpha reliability coefficient |
| maxwell | Maxwell's RE coefficient for binary data |
| meancor | Mean of bivariate correlations between raters |
| meanrho | Mean of bivariate rank correlations between raters |
| N.cohen.kappa | Sample Size Calculation for Cohen's Kappa Statistic |
| N2.cohen.kappa | Sample Size Calculation for Cohen's Kappa Statistic with more than one category |
| print.icclist | Default printing function for ICC results |
| print.irrlist | Default printing function for various coefficients of interrater reliability |
| rater.bias | Coefficient of rater bias |
| relInterIntra | Inter- and intra-rater reliability |
| robinson | Robinson's A |
| stuart.maxwell.mh | Stuart-Maxwell coefficient of concordance for two raters |
| video | Different raters judging the credibility of videotaped testimonies |
| vision | Eye-testing case records |