Module 4: Analyzing image content with computer vision
AI-aided content analysis of sustainability communication
Lesson 4.3: Interpreting the results of CV analysis
lecture slides
lecture video
lecture text
Interpreting results of CV analysis
Interpreting computer-vision (CV) outputs is the final step of the image-analysis workflow and should explicitly answer the research questions you posed at the outset. Although conceptually similar to interpreting NLP results, here the evidence is visual rather than textual: labels, counts, bounding boxes, and spatial patterns replace tokens and n-grams. Because NLP is often more mature in the social sciences, it can serve as an interpretive template (e.g., map “object frequency” to “term frequency,” or “salience in the frame” to “prominence in a paragraph”), but CV frequently demands more “out-of-the-box” reasoning about composition, color, scale, and co-presence in the scene. The key is to connect model outputs (e.g., class probabilities, detection counts) to constructs in your theory (e.g., risk framing, solution framing, corporate presence) and to articulate how the visual evidence supports or disconfirms your hypotheses. Throughout, report uncertainty (confidence scores, error patterns) and discuss alternative explanations to ensure your claims remain tied to the questions you set out to answer.
Operationalizations using image features
Operationalizing communication concepts in CV means translating them into measurable variables derived from images. Define your dependent variable clearly (e.g., the probability or count of “nature imagery,” “industrial infrastructure,” or “logo presence”) and specify how it is measured (classification label, detection count, or segmented area share). Then define explanatory variables—often categorical—such as organization, sector, campaign, channel, or time period, and code them consistently (e.g., one-hot or hierarchical groupings). State expectations ex ante (e.g., organizations with lower environmental impact will display higher “nature imagery” prevalence), and link each expectation to a specific statistical test or model. This design anchors interpretation, reduces post-hoc bias, and makes it clear which visual features are intended to serve as indicators of the underlying communication constructs.
Comparisons across organizations
When comparing organizations with high versus low environmental impact, specify what you expect to see visually and why. High-impact firms may feature mitigation and infrastructure cues (plants, pipelines, protective gear, dashboards), whereas lower-impact firms may emphasize ecosystems, communities, and everyday practices (forests, biodiversity, people in natural settings). Note that overlap is possible—both groups might depict wind turbines or lab settings—so your analysis should distinguish distinctive features (used disproportionately) from common ones (shared baseline imagery). Use stratified summaries and normalization (e.g., per 100 images) to ensure fair comparisons across unequal sample sizes. Finally, contextualize differences: are they consistent across channels and time, or limited to specific campaigns and events?
Summarizing results of image analysis
Start by reading the image-classification dataframe and ensuring labels and metadata are tidy and interpretable (splitting long, compound class labels into meaningful tokens can aid grouping and display). Summarize key metrics—class frequencies, detection counts per image, mean confidence—before fitting models that test associations (e.g., logistic or Poisson regression, mixed effects for campaigns). Report effect sizes with uncertainty, and use appropriate significance controls when testing many classes. Clarify whether your stance is exploratory (pattern finding, hypothesis generation) or confirmatory (pre-specified hypotheses), since this affects how results should be weighed. Well-structured summaries turn raw model outputs into evidence that answers substantive questions.
Select, filter, aggregate
Prior to modeling, select the variables that correspond to your conceptual framework (dependent and independent), and filter rows to remove noise: nulls, corrupt images, or predictions below class-specific confidence thresholds. Aggregate with simple, interpretable functions—counts, proportions, means—at analysis-relevant levels (image, campaign, organization, time window). Build compact summary tables (e.g., class-by-organization with normalized proportions) that feed directly into statistical tests and figures. This disciplined select-filter-aggregate routine reduces variance from noisy predictions and makes downstream interpretation more robust and transparent.
Visualizing results of image analysis
Use standard numeric visualizations to communicate prevalence and differences (bar charts, ridgeline densities, heatmaps of feature co-occurrence), and complement them with CV-specific visuals that show what the model saw (detections drawn as bounding boxes, segmentation overlays). Include diagnostic views—confusion matrices, precision–recall curves, examples of edge cases and misclassifications—to help audiences gauge reliability. Distinguish between data visualizations (what appears in the corpus) and method visualizations (how the model behaves), and use each for its purpose. When showing trends, be explicit about normalization and uncertainty so visual comparisons map cleanly to your claims.
Grouped bar plots
Grouped (clustered) bar plots are a compact way to display multi-dimensional comparisons while retaining the simplicity of bivariate bars. Use groups (color/hue and legend) to encode organizations and bars to encode image classes or themes, normalizing to proportions to control for different sample sizes. This format highlights both overlapping labels (common imagery across organizations) and distinctive labels (classes over-represented in one group). Order bars by effect size or prevalence, add error bars or CIs where appropriate, and keep legends concise. Well-designed grouped bars make it easy to see where visual narratives converge and where they diverge.