Module 4: Analyzing image content with computer vision


Lesson 4.3: Interpreting the results of CV analysis

AI-aided content analysis of sustainability communication

nils.holmberg@iko.lu.se

Interpreting results of CV analysis

  • Interpretation is the final step that must explicitly answer the research questions you posed at the outset.
  • In CV the evidence is visual—labels, counts, boxes, and spatial patterns—rather than textual tokens and n-grams.
  • NLP can serve as a template for analogies, but CV also demands reasoning about composition, color, scale, and co-presence.
  • Connect model outputs such as class probabilities and detection counts to theoretical constructs like risk or solution framing.
  • Report uncertainty and alternative explanations so claims remain proportional to confidence and grounded in the study design.

Operationalizations using image features

  • Translate communication concepts into measurable variables derived from images.
  • Define a clear dependent variable and its measurement as a label, count, or segmented area share.
  • Specify explanatory variables, often categorical, and code them consistently across organization, sector, campaign, and time.
  • State a priori expectations and link each to a specific statistical test or model.
  • Use this design to limit post-hoc bias and to clarify which features indicate the constructs of interest.

Comparisons across organizations

  • Begin with theory-driven expectations about how visuals should differ between organizations.
  • High-impact firms are expected to feature mitigation and infrastructure cues.
  • Low-impact firms are expected to emphasize ecosystems, communities, and everyday practices.
  • Distinguish common imagery from distinctive features by comparing normalized rates rather than raw counts.
  • Contextualize differences across channels and time to avoid attributing one-off campaigns to enduring strategies.

Summarizing results of image analysis

  • Tidy the classification dataframe and make labels interpretable, including splitting compound class names.
  • Produce core summaries such as class frequencies, detections per image, and mean confidence.
  • Fit models that test associations, for example logistic or Poisson regression with campaign random effects.
  • Report effect sizes with uncertainty and control for multiple comparisons when many classes are tested.
  • Declare whether the analysis is exploratory or confirmatory so readers weigh results appropriately.

Select, filter, aggregate

  • Select dependent and independent variables that reflect the conceptual framework.
  • Filter out nulls, corrupt items, and predictions below class-specific confidence thresholds.
  • Aggregate with simple functions—counts, proportions, and means—at image, campaign, organization, or time levels.
  • Build compact summary tables such as class-by-organization with normalized proportions.
  • Use this disciplined routine to stabilize estimates and make interpretation transparent and reproducible.

Visualizing results of image analysis

  • Use numeric graphics such as bar charts, ridgeline densities, and co-occurrence heatmaps to show prevalence and differences.
  • Complement them with CV-specific visuals that reveal what the model saw, including bounding boxes and segmentation overlays.
  • Include diagnostic views such as confusion matrices, precision–recall curves, and curated misclassification examples.
  • Separate data visualizations that describe the corpus from method visualizations that describe model behavior.
  • Display trends with clear normalization and uncertainty so visual comparisons map cleanly to the claims.

Grouped bar plots

  • Choose grouped bars to visualize multi-dimensional comparisons while retaining simple bivariate structure.
  • Encode organizations as groups and image classes or themes as bars within each group.
  • Normalize to proportions to control for unequal sample sizes across organizations.
  • Order bars by prevalence or effect size and add error bars or confidence intervals where appropriate.
  • Highlight where labels overlap across organizations and where distinctive imagery is over-represented.