Plot inter-rater agreement diagnostics using base R
Source:R/api-plotting.R
plot_interrater_agreement.RdPlot inter-rater agreement diagnostics using base R
Arguments
- x
Output from
fit_mfrm()orinterrater_agreement_table().- diagnostics
Optional output from
diagnose_mfrm()whenxismfrm_fit.- rater_facet
Name of the rater facet when
xismfrm_fit.- context_facets
Optional context facets when
xismfrm_fit.- exact_warn
Warning threshold for exact agreement.
- corr_warn
Warning threshold for pairwise correlation.
- plot_type
"exact","corr", or"difference".- top_n
Maximum pairs displayed for bar-style plots.
- main
Optional custom plot title.
- palette
Optional named color overrides (
ok,flag,expected).- label_angle
X-axis label angle for bar-style plots.
- preset
Visual preset (
"standard","publication", or"compact").- draw
If
TRUE, draw with base graphics.
Details
Inter-rater agreement plots summarize pairwise consistency for a chosen rater facet. Agreement statistics are computed over observations that share the same person and context-facet levels, ensuring that comparisons reflect identical rating targets.
Exact agreement is the proportion of matched observations where both raters assigned the same category score. The expected agreement line shows the proportion expected by chance given each rater's marginal category distribution, providing a baseline.
Pairwise correlation is the Pearson correlation between scores assigned by each rater pair on matched observations.
The difference plot decomposes disagreement into systematic bias (mean signed difference on x-axis: positive = Rater 1 more severe) and total inconsistency (mean absolute difference on y-axis). Points near the origin indicate both low bias and low inconsistency.
The context_facets parameter specifies which facets define "the
same rating target" (e.g., Criterion). When NULL, all non-rater
facets are used as context.
Plot types
"exact"(default)Bar chart of exact agreement proportion by rater pair. Expected agreement overlaid as connected circles. Horizontal reference line at
exact_warn. Bars colored red when observed agreement falls below the warning threshold."corr"Bar chart of pairwise Pearson correlation by rater pair. Reference line at
corr_warn. Ordered by correlation (lowest first). Low correlations suggest inconsistent rank ordering of persons between raters."difference"Scatter plot. X-axis: mean signed score difference (Rater 1 \(-\) Rater 2); positive values indicate Rater 1 is more severe. Y-axis: mean absolute difference (overall disagreement magnitude). Points colored red when flagged. Vertical reference at 0.
Interpreting output
Pairs below exact_warn and/or corr_warn should be prioritized for
rater calibration review. On the difference plot, points far from the
origin along the x-axis indicate systematic bias; points high on the
y-axis indicate large inconsistency regardless of direction.
Typical workflow
Select rater facet and run
"exact"view.Confirm with
"corr"view.Use
"difference"to inspect directional disagreement.
Further guidance
For a plot-selection guide and a longer walkthrough, see
mfrmr_visual_diagnostics and
vignette("mfrmr-visual-diagnostics", package = "mfrmr").
Examples
toy <- load_mfrmr_data("example_core")
fit <- fit_mfrm(toy, "Person", c("Rater", "Criterion"), "Score", method = "JML", maxit = 25)
p <- plot_interrater_agreement(fit, rater_facet = "Rater", draw = FALSE)
if (interactive()) {
plot_interrater_agreement(
fit,
rater_facet = "Rater",
draw = TRUE,
plot_type = "exact",
main = "Inter-rater Agreement (Customized)",
palette = c(ok = "#2b8cbe", flag = "#cb181d"),
label_angle = 45,
preset = "publication"
)
}