Skip to main content
  • Regular article
  • Open access
  • Published:

A large scale study of reader interactions with images on Wikipedia

Abstract

Wikipedia is the largest source of free encyclopedic knowledge and one of the most visited sites on the Web. To increase reader understanding of the article, Wikipedia editors add images within the text of the article’s body. However, despite their widespread usage on web platforms and the huge volume of visual content on Wikipedia, little is known about the importance of images in the context of free knowledge environments. To bridge this gap, we collect data about English Wikipedia reader interactions with images during one month and perform the first large-scale analysis of how interactions with images happen on Wikipedia. First, we quantify the overall engagement with images, finding that one in 29 pageviews results in a click on at least one image, one order of magnitude higher than interactions with other types of article content. Second, we study what factors associate with image engagement and observe that clicks on images occur more often in shorter articles and articles about visual arts or transports and biographies of less well-known people. Third, we look at interactions with Wikipedia article previews and find that images help support reader information need when navigating through the site, especially for more popular pages. The findings in this study deepen our understanding of the role of images for free knowledge and provide a guide for Wikipedia editors and web user communities to enrich the world’s largest source of encyclopedic knowledge.

1 Introduction

Almost 20 years after its birth, Wikipedia has become the reference for online diffusion of free encyclopedic knowledge, reaching 54M articles in 313 language editions.Footnote 1 Its content is generated by the collaborative effort of a large community of editors, and provides a reliable source of information for web users [1]. Knowledge on Wikipedia is mainly conveyed under the form of written text, but also through other types of content, such as references and images.

The space of visual content on Wikipedia is vast. English Wikipedia alone contains more than 5M distinct images, the majority of which is hosted by Wikipedia’s sister project Wikimedia Commons,Footnote 2 the world’s largest free visual knowledge repository. The genesis of Wikipedia images involves not only the contribution of Wikipedia editors, but also the participation of visual content creators. Visual content on Commons and Wikipedia originates from individual photographers, artists, web users and cultural institutions in the GLAM spaces,Footnote 3 who actively release their works of art for free and public use.

Given the crucial role of Wikipedia as a central hub for knowledge sharing and learning, understanding how images are used on Wikipedia is particularly important. A vast body of literature in experimental psychology has shown the impact of images for learning and engaging with knowledge. Images positively affect comprehension and increase attention on the textual material [2]. Despite their importance, while other aspects of Wikipedia have been widely studied [35], little is known about visual content and its usage, with only a few studies looking at cross-language image diversity [6], and the communities of Wikipedia “image” editors [7].

In this paper, we fill this gap in the literature by providing for the first time a comprehensive overview of how readers interact with images in (English) Wikipedia. We quantify and characterize reader engagement with images when browsing the encyclopedia using traffic data and we explore the role played by images in the exploration of free knowledge. To operationalize reader engagement, we adopt the most widely-used metrics in web user studies [8, 9]: we compute click-through rate on images, and conversion rate on illustrated and unillustrated page previews. While only partially representing the complex, multifaceted notion of interest [10], these implicit signals do reflect an expression of engagement with visual content and they provide a solid baseline for an initial overview of readers’ interactions with images. More specifically, we address three major research questions:

RQ1::

To what extent are readers interacting with images on Wikipedia? And what is the relation with engagement values on other types of content?

RQ2::

What drives reader’s engagement with images when reading Wikipedia articles? What are the visual and contextual factors that influence image interactions?

RQ3::

Do images support reader’s need for additional information when navigating Wikipedia? Are images helpful to delve into contextual information provided by the article?

In addressing these questions, we make the following contributions:

  • RQ1: We collect a large dataset of reader interactions with images in English Wikipedia over one month and characterize the landscape of Wikipedia images with several features inspired by experimental psychology and web user studies (Sect. 4.3). We quantify reader engagement with images and find that, on average, readers click with images 1 in every 29 pageviews on English Wikipedia, ten times more often than with references (RQ1, Sect. 5).

  • RQ2: To visualize the factors impacting reader engagement with Wikipedia images, we perform a set of multivariate analyses on the image features extracted and find that readers interact more often with images of monuments, maps, vehicles, and unfamiliar faces (RQ2, Sect. 6).

  • RQ3: To understand whether images support readers’ need for additional contextual information when navigating Wikipedia, we design a matched observational study based on page previews, i.e., the short article summaries that are displayed when users hover on links to other Wikipedia pages (RQ3, Sect. 7). We find a negative effect of the presence of images on the proportion of articles’ page previews that convert into a visualization of the full article page.

We conclude (Sect. 8) that the visual preferences of Wikipedia readers are radically different compared to web users in photo-sharing platforms or image search engines, where images of people and celebrities largely predominate. We also find that images on Wikipedia appear to fulfill part of the cognitive function typical of illustrations in instructional settings supporting readers’ information need. Finally, we discuss theoretical implication of this research and its important repercussions on how the Wikipedia communities organize and prioritize the inclusion of visual content and how the broader web and content creators could contribute to the web with free visual knowledge.

2 Related work

Our work is highly related to research from experimental psychology, computer vision, information retrieval, and computational social science, looking at how readers navigate knowledge.

2.1 The role of text illustrations

A substantial body of literature from experimental and educational psychology studied the role of images for knowledge understanding and learning. Researchers have found that, very often, images in association with text help to support learning in instructional contexts [11], also in online settings [1215], especially when images are carefully curated, described, and positioned in the text [16, 17]. Beyond this cognitive purpose of facilitating content comprehension and providing complementary information, textual illustrations can have many other functions: the attentional function, meaning that images can help to attract attention to the information in the textual form; the affective function—images help enhance emotions and enjoyment when reading a text; and the compensatory purpose of supporting poor readers [2]. While testing the role of images for knowledge understanding is beyond the scope of this paper, we borrow some ideas from these works to analyze reader interactions with images and design features and experiments aimed at replicating some of their findings.

2.2 Image interestingness

Several studies in computer science have looked at what makes images interesting from a computational perspective. Researchers have typically described interestingness in two ways. Visual interestingness is the extent to which an image can hold or catch the viewer’s attention due to its intrinsic visual qualities (see Constantin et al. [10] for a review of the most recent works in this space). Researchers have found that, for example, images are more interesting when they are more aesthetically pleasing or when their content is visually complex or unfamiliar [18]. Social interestingness is often also called popularity and corresponds to the extent to which an image is liked by a large number of people in a community. Social interestingness depends on the social dynamics of the platforms where images are shared, the pictures visual content [8, 19, 20] and the text associated with them [21]. Most of these previous works focus on predicting image popularity in photo sharing platforms such as Flickr [19, 21] or Instagram [8], specifically designed to increase social interestingness in images. Unlike these existing works, we analyze here for the first time how readers engage with images in the context of online free knowledge spaces. We model the complex interplay between encyclopedic knowledge, pictorial representations, and reader engagement and explore the role of images informational support for Wikipedia articles.

2.3 Image search behavior

Related works have investigated web user behavior in image search engines. Researchers have found that, in general, the most popular queries in image search engines are about people, celebrities, and entertainment [2224]. By comparing image search behavior with text search behavior, several studies have found that image search sessions are heavier in interaction and exploration than the more “focused” textual sessions [25], although, in a later study, O’Hare et al. found that web image search behavior is nonuniform across query types [9]. While the scope of this work is different from this body of research, we will factor into our analysis findings from this area.

2.4 Images on Wikipedia

Recent works have quantified the monetary value and underlined the social contribution of Wikipedia’s visual side [26, 27]. However, despite their important role, the space of images on the encyclopedia has rarely been investigated. Given the richness of their semantics, researchers have worked on building structured datasets from images in Wikimedia Commons, and Wikipedia [28, 29]. However, only a few works have focused on understanding editing behavior concerning images. A seminal study looked at understanding communities of editors who curate the visual content of the encyclopedia [7]. More recently, He et al. [6] measured the visual diversity of Wikipedia, finding that cross-language image diversity is higher than the diversity of textual content and that many images are unique to specific language editions. Moreover, Navarrete et al. [30] investigated the role of image paintings on Wikipedia, finding that they are extensively used to illustrate also non-art-related topic and that their audience is even larger than that of art-related articles. While most of these works focus on the image content or the Wikipedia editor communities, we study here the complementary aspect of how readers interact with the visual content.

2.5 Studying Wikipedia readers

A few studies have focused on Wikipedia reader behavior, including reader article topic preferences [31, 32], reader perception of the site performances [33], or reader informational need [4, 34]. More recently, Piccardi et al. worked on quantifying reader engagement with citations [35] and external links [36]. The authors collected a large dataset of reader interactions with footnotes and references and showed that only a tiny portion of readers engage with citations on Wikipedia. In this direction, this work gives an additional perspective on Wikipedia readers’ behavior, focusing on the volume and characteristics of reader interactions with visual content.

2.6 User engagement metrics

To quantify user engagement with images when reading Wikipedia, we borrow metrics used by several studies in the computational advertising field and user engagement studies [31]. While these works aim to predict engagement metrics such as conversion rate [37, 38], namely the percentage of landing page visits that result in a target action, or the click-through rate [3941], namely the ratio between clicks and impressions, we use these metrics here as a means to decode Wikipedia reader behavior.

3 Images on Wikipedia

Images are a core component to help readers interpret knowledge in the encyclopedia and complement the textual information on Wikipedia articles. As English Wikipedia guidelines put it, “The purpose of an image is to increase reader understanding of the article’s subject matter, usually by directly depicting people, things, activities, and concepts described in the article.”Footnote 4

Images are added to Wikipedia by hundreds of thousands of editors from all around the world, following a Manual of Style maintained by the Wikipedia community.Footnote 5 In essence, images added to Wikipedia articles have to be relevant to the article’s content and of high photographic quality. The majority of images in the encyclopedia are hosted in Wikimedia Commons, the largest free visual knowledge repository. Images in Wikimedia Commons must be either of the public domain or licensed under a free license allowing anyone to reuse the material for any purpose, including commercial purposes.

Images can be placed in different parts of an article (see Fig. 1 and 2). Readers can find images in the infobox, a table summarizing the main facts about the article’s subject, usually placed in the top-right corner of the page on desktop browsers or at the top of the screen in mobile browsers. Images can also be added inline, namely individually near the relevant text in the article body. Finally, when images are too many to be placed within the text body, they can be collected into galleries, generally added at the bottom of the articles. On desktop devices, images are also available in article previews, i.e., the pop-up containing the article summary and an image (when available) which gets displayed whenever a reader hovers over on a link to another Wikipedia article.

Figure 1
figure 1

Examples of the two types of images visualization that we investigate in this study. The Media Viewer is opened when images are clicked, while Page Previews are shown when the reader hovers over a link to another page

Figure 2
figure 2

Examples of the three types of positions (infobox, inline, and gallery) of images within a Wikipedia article

For readers and editors who are interested in exploring Wikipedia visual content in further detail, images in articles are clickable: when clicked, an image is previewed in a visualization tool called Media Viewer.Footnote 6 The Media Viewer overlays on the article and displays the image in a larger size, and additional metadata below.

But how much visual content is available for readers to explore? If we take English Wikipedia, the largest language edition of the online encyclopedia, as of March 2021, it counted 6.2M articles, for a total of 5M unique images. As many other quantities on the Web, the distribution of images across Wikipedia articles follows a power law: as shown in Fig. 3(A), only \(\approx 44\%\) of pages in English Wikipedia are illustrated.

Figure 3
figure 3

Cumulative distributions of (A) number of images per article and (B) the number of articles per image

One reason for the number of missing images is the effort needed to illustrate articles. Wikipedia editors need to find the right image match for an article by searching through millions of images in Wikimedia Commons. However, when relevant images are not present in Commons, editors will have to search other sources. First, the right pictures for an article need to exist somewhere on the Web (or in the world): otherwise, someone, Wikipedia editors, photographers, GLAM institutions, or other users, must create or retrieve them. Second, images have to be free to reuse. If images are not free-licensed, editors’ and authors’ efforts will be needed to make them publicly available. Only then images can be hosted in Wikimedia Commons and finally added to Wikipedia articles. To help with these efforts, the Wikimedia movement organizes several initiatives, e.g., Wiki Loves Monuments, encouraging photographers to add free images of monuments,Footnote 7 or the #WPWP campaign, which helps editors add images to unillustrated Wikipedia articles.Footnote 8

Given the central role of Wikipedia and its diverse content nature, knowing whether and how readers use visual information could help prioritize efforts around the visual enrichment of Wikipedia.

4 Data collection and methods

To answer our research questions, we first need to estimate the volume of Wikipedia articles and their images, collect data about reader interactions with those, and characterize them through feature extraction.

4.1 Collecting article and image counts data

To measure the number of articles and images, we used the HTML version of English Wikipedia at the end of March 2021. We collected 6.2M documents, and we parsed them to extract the images’ URLs, caption text, resolution, and position on the page. Using the CSS class in the HTML code, we exclude all images that appear as icons (for example, portals or Wikiprojects). Additionally, for each page, we also record the article length as the number of characters.

Out of the 6.2M articles, 2.7M (44%) contained at least one image, for a total of 5M unique images across all English Wikipedia articles. The vast majority of the articles (91%) contain two images or less, while only 1.5% has more than eight images (see Fig. 3(A)). On average, there are 2.3 images per illustrated article. Around 84% of images is unique to the article where it appears, while 16% of the images appear in more than one article (see Fig. 3(B)).

4.2 Collecting article and image traffic data

We obtained the reader interactions with images for desktop and mobile browsers by processing the server access logsFootnote 9 collected from 1st to 28th of March 2021. We restricted our analysis to only human interactions by ignoring traffic from bots thanks to a set of heuristics developed by Wikimedia’s Analytics team.Footnote 10 For privacy reasons, we worked with an anonymized version without sensitive information. Since the logs do not contain any explicit identifier for the user, before the anonymization, we assigned a random id based on IP and user-agent similar to previous work [42]. In addition, we discarded all the events coming from logged-in users, the events of any user that edited a page, and the events originated from countries where not all days have more than 500 pageviews consistently. This filtering ensures more privacy for the Wikipedia readers by dropping around 3% of the data.

Over the considered period, we selected from the web logs all requests that reflect three types of actions:

  • Imageviews: these requests correspond to image visualizations in the Media Viewer after a user clicks on an article image.

  • Pageviews: these are requests logged every time a user visits a Wikipedia page. For the scope of this study, we select only pageviews of articles with at least one image.

  • Page previews: these requests are logged whenever a user hovers over a link to an article. To remove the effect of casually generated page previews, we only keep those previews that are shown for at least one second. Note that page previews are generated only on desktop devices.

We aggregated these image-related events at the user level by using the previously assigned id to obtain sorted sequences of actions from the same user, which we refer to as sessions.

In our analysis, we do not consider exogenous time-dependent events’ impact and the role that external image search engines may play in directing users to Wikipedia. For these reasons, we filter out all the incoming traffic generated from Google Image Search, which represents by far the most used image retrieval engine from which people access Wikipedia’s visual content. Nevertheless, the pageviews originating from Google Image search account for 0.006% of the total, making their impact negligible.

In our data collection, we extracted interactions for 1.5B sessions. In Fig. 4 we report the distributions of sessions by the number of imageviews, pageviews, and previews. The distributions are heavily skewed, with 91% and 94% of sessions having less than 10 pageviews on desktop and mobile devices, respectively, and 99% of sessions having less than 10 imageviews both on desktop and mobile devices during our data collection period. Similarly, 79% of sessions have generated less than 10 previews. Users with extensive sessions (i.e. “power users”), that may be over-represented, are therefore limited in our analysis. Over one month, 100% of the illustrated articles have been loaded at least once, accounting for a total of 7.1B pageviews, 461M imageviews, and 49M previews events in our dataset. We find that most pageviews are generated from mobile devices (59% from the mobile site), while most imageviews are generated from desktop (58% from desktop).

Figure 4
figure 4

Cumulative distributions of number of sessions by (A) imageviews, (B) pageviews, and (C) previews partitioned by desktop (in blue) and mobile device (in orange). Page previews are available only on desktop devices

4.3 Mining image content and context

To investigate the factors that make images engaging when reading Wikipedia, we characterize the pictures in our dataset with several features related to the visual context and content. Our choice of features is largely inspired by the literature around the cognitive perception of images in instructional or web environments.

4.3.1 Contextual factors

Images on Wikipedia are not isolated items. Instead, they exist in context, providing epistemic support to the article they are illustrating. To extract features from the image context, we resort to previous literature on Wikipedia reader behavior and experimental psychology studies on the role of images in instructional settings. Note that 16% of the images appear in multiple articles. Since the same image may appear in very different articles, thus belonging to very different contexts, we treat such images as distinct.

Page topic

In a previous study, Piccardi et al. [35] found that Wikipedia reader engagement with references varies with the article topic. To test whether the reader’s need for visual support similarly varies across subject matters, we extract, for each page in our dataset, a topic vector, using the Wikidata topic model.Footnote 11 The classifier takes as input the Wikidata item of a Wikipedia page, and it returns a 64-dimensional vector containing the probability that the article belongs to the topics of the Wikiproject hierarchy.Footnote 12 To reduce the dimensionality of the topic vector, we consider the second level of the topic taxonomy accounting for 31 topics. We then rearranged some of the topics into coarse-grained topics, namely media, internet culture, and performing arts into entertainment, chemistry and biology into biology, computing and libraries & information into computer science, mathematics and physics into maths & physics. Figure 5(A) shows the distribution of images by article topic. Geographic articles are the most illustrated, containing \(1/4\) of the images in our dataset. Biographies, making up 30% of the articles on Wikipedia, also contain around 15% of the images. Topics such as entertainment (movies, plays, books), visual arts, transportation, military, biology, and sports follow, covering together another third of the images in English Wikipedia. A summary of the numerical values can be found in the Additional file 1 (Supplementary Table).

Figure 5
figure 5

(A) Fraction of images by topic (in blue) and fraction of images with faces (in orange). (B) Image-specific CTR by article topic

Page length

One of the possible functions of text illustrations in learning contexts is to enrich and complement the textual content with additional material [2]. To investigate to which extent images are used to complement the lack of textual information, we measure the textual richness as the length of each article in characters. The distribution of the number of images by text length as shown in Fig. 6 is log-normal, with most images in English Wikipedia being found in articles between 1k and 100k characters long.

Figure 6
figure 6

Feature distributions. Spearman’s rank correlation coefficient ρ between the numerical features and the iCTR at the top of each panel (\(p<0.001\) for each feature)

Page popularity

Previous work analyzing reader behavior with respect to Wikipedia citations [35] found that there is an inverse relation between article popularity and reference click-through rate. To test whether this relation is valid also in the case of interactions with images, we compute a page popularity feature for each image by computing the total monthly pageviews for the page where an image appears. As in Piccardi et al. [35], the page popularity follows a power-law distribution (Fig. 7), with 70% images having average monthly pageviews in the range between 50 and 10k.

Figure 7
figure 7

Distributions of (A) values of iCTR by page popularity partitioned by device and (B) number of images per page popularity. Spearman’s rank correlation coefficient \(\rho _{iCTR}\) between iCTR and pageviews in the inset (\(p<0.001\)). The axes are in log scale

Readability

Although not completely verified, another function of images in textual knowledge is to facilitate text comprehension, especially in the case of reading difficulties [2]. To take into account this function in our study, we quantify the reading ease by computing the Flesch readbility score, reflecting the “comprehension difficulty of written material” [43], on the text of each article in our dataset. We compute the readability score for all the pages containing an image, and plot the resulting distribution in Fig. 6: most of the images on Wikipedia are in articles detected as “Fairly difficult to read” (score 50–60), or “Difficult to read” (score 30–50).

Length of the image caption

Studies in educational technologies have found that the usage of captions marginally enhances the usefulness of text illustrations [17]. To operationalize the presence of captions as a contextual feature of the images in our dataset, we store the average number of words used to caption each image when appearing in a Wikipedia article. We can see from Fig. 6 the caption length following a Tweedie distribution with a large fraction of the images without a description and the majority of existing captions centered around ten words.

Image placement

How images are placed in the text can play a crucial role in the knowledge exploration experience [16], and researchers investigating Wikipedia reader behavior showed that people tend to engage more with content (in this case, internal hyperlinks) which lies at the top of the article [44]. At the same time, Wikipedia editors follow specific placement guidelines when illustrating an article. To investigate the role of image placement on Wikipedia article consumption, we extract the image’s text offset, i.e., the relative position of the image with respect to the length of the article, as well as the image position, a categorical feature which can take the values \(\{\mathit{infobox}, \mathit{inline}, \mathit{gallery}\}\), depending on the template used to add the image to the article. From the plots in Fig. 6 we can see that only 36% of the images in our dataset is generally placed in infoboxes, while only 16% can be found in galleries, and that the majority of inline images are generally placed at the top of the article (see offset). A summary of the numerical values can be found in the Additional file 1 (Supplementary Table).

Image resolution

In addition to their position, the viewer’s attention may also be driven by the size of an image. According to the Wikipedia’s Image Size guidelines,Footnote 13 editors should choose the appropriate image size in proportion of its level of details. However, readers may still tend to click on small images that are inherently difficult to observe. To investigate the role of the image size, we compute the image resolution in pixels for each image. As shown in Fig. 6, image resolutions vary across different scales, mostly ranging from 10k to 100k pixels.

4.3.2 Visual factors

The content of pictures plays a key role in driving readers’ attention to both the images [18] and the text on the page [2]. To understand the type of visuals that elicit higher levels of interactions with Wikipedia images, we run a set of computer vision-based classifiers. Since training a classifier to detect every concept in Wikipedia’s visual knowledge would be practically infeasible, we instead focus on three main indicators, based on extensive literature from visual and social interestingness prediction.

Image quality

Visual aesthetics, or image quality, is one of the top visual factors driving the viewer’s attention to an image [18]. At the same time, researchers have found that not all images which receive much attention from web communities are actually of high quality [45], and that a lot of socially uninteresting pictures are very beautiful. We investigate here whether the quality of an image plays an important role in eliciting Wikipedia reader attention. To do so, we design a Wikipedia Image Quality classifier, as follows.

  • We collect a training set of images annotated with a binary (high/low) image quality score. To annotate images, we resort to the highly curated categories that Wikimedia Commons editors assign to images. We download \(141{,}984\) images from the Quality images category from Commons:Footnote 14 these are high-quality images that have to meet Commons’ quality guidelinesFootnote 15 before being voted and promoted as Quality images by the community through a highly selective process. Only a few images make it to the “image quality” category: there is, therefore, a large consensus on the quality of the images in that category. To collect low-quality images, we simply randomly sample an approximately equal number of pictures (\(169{,}310\)) from the large pool of Commons images. These are very likely to be low quality, as images randomly drawn from Commons tend to have a small resolution, and they are rarely used to illustrate Wikipedia articles [27].

  • We next train a deep neural network using transfer learning: we fine-tune a pre-trained model, originally designed to classify image objects, using the image quality data collected. We use the Inception-v3 [46] deep network pre-trained on the 1000-classes ImageNet dataset [47], as it was proved to be a good starting dataset for transfer learning tasks [48]. We use 90% of the data for training and the rest for validation, and we train the last layer of the network over 10,000 iterations with the data collected. The fine-tuned classifier achieves 77% accuracy on a balanced test set.

The resulting image quality classifier, given any image, outputs a quality score in the range \([0,1]\) which corresponds to the probability that the image belongs to the “High Quality” class. As shown in Fig. 6, most images in our dataset have a very low-quality score.

Presence of faces

In line with several studies showing the importance of faces for web users’ positive reactions and engagement with images [8, 49], we also extract information about the presence of faces of people in the image. We use MTCNN [50] to detect faces and their bounding box in an image. For a given image, we then output a binary feature indicating whether it contains at least one face or not. We find that around \(1/3\) of the images on Wikipedia have at least one face (see Fig. 6), and most of those are in articles about biographies, entertainment, and sports (see Fig. 5(A)).

Outdoor setting

Literature around image interestingness and aesthetics [18] has shown that outdoor images tend to elicit the viewer’s interest more than indoor images do. To extract the information about the image scene setting, we use a Wide Residual Network [51] trained on MIT’s Places [52], an image dataset with 10M images annotated with 365 scene types, and indoor/outdoor labels. This classifier, given an image, outputs an outdoor score which reflects the probability of the image being an outdoor scenery. When the feature is ≤0.5, the image is likely to be an indoor scenery. In our dataset, indoor and outdoor images are almost equally distributed, with a slight prevalence of outdoor pictures.

4.4 Engagement metrics

To quantify the volume of readers’ interactions with visual content, we introduce the following metrics:

Global click-through rate

The global click-through rate (gCTR) measures the overall reader engagement with images. It is defined as the fraction of reading sessions with at least one interaction with an image. Formally, for each session s, let \(C(s,p)\) be the indicator function that is 1 if at least one image was clicked on page p by the respective reader, and 0 otherwise. Moreover, let \(N(p)\) be the number of distinct reading sessions during which page p was loaded. We define the global click-through rate as

$$\begin{aligned} \mathrm{gCTR} = \frac{\sum_{s} \sum_{p} C(s,p)}{\sum_{p} N(p)}, \end{aligned}$$
(1)

where p ranges over the set of pages that contain at least one image.

Image-specific click-through rate

The image-specific click-through rate (iCTR) measures how much engagement a Wikipedia image elicits. It is defined as the ratio of clicks to impressions. Formally, let \(N(i)\) be the number of distinct sessions with clicks on image i and \(N(p_{i},i)\) the number of distinct sessions that viewed page \(p_{i}\) where the image is placed, the image-specific click-through rate is

$$\begin{aligned} \mathrm{iCTR}(i) = \frac{N(i)}{\sum_{p_{i} \in P_{i}} N(p_{i},i)}, \end{aligned}$$
(2)

where \(p_{i}\) ranges over the set \(P_{i}\) of pages containing i.

Conversion rate

The conversion rate (CR) quantifies the probability of clicking on an article link after its preview is shown in another article. Formally, for each page p and session s, we denote by \({C(s,p)}\) the indicator function that is one if session s has clicked on a link to page p after seeing its preview. Moreover, we denote by \(N(p)\) the total number of distinct sessions that loaded a preview of p. The conversion rate for page p can be written as:

$$\begin{aligned} \mathrm{CR}(p) = \frac{\sum_{s} C(s,p)}{N(p)} \end{aligned}$$
(3)

In the following sections, we restrict our analyses to images visualized by at least 50 readers during the period of our data collection in order to reduce the effect of rarely viewed articles and obtain a reliable estimate of the quantities above. This results in a set of 3.2M unique images displayed in 2.7M articles.

5 RQ1: to what extent are readers interacting with images in Wikipedia?

The first step of our analysis is to quantify the volume of readers’ interactions towards visual content when reading Wikipedia. To this aim, we compute the global click-through rate and image-specific click-through rate on our data and find the following.

5.1 Overall engagement with images: the global click-through rate

We find that the gCTR across all pages in English Wikipedia with at least one image is 3.5%, meaning that around 3.5 out of 100 times readers visit a page, they also click on an image. This metric is higher for desktop (5.0%) and lower for mobile web users (2.6%), probably due to differences in the way readers navigate Wikipedia on the two devices and the better Media Viewer experience on desktop. Over time, the behavior also changes depending on the device used. For example, on desktop, readers tend to click more often on images during weekdays (Monday to Friday), with an increase of 5.5% over weekends. However, on mobile, there is no significant difference between week and weekends. To understand whether these values represent a high or low level of engagement, we can compare them with engagement metrics on another type of article content, namely article’s references. According to Piccardi et al. [35], the gCTR on citations in English Wikipedia is 0.29%, thus around ten times lower than for images. This observation suggests that images tend to elicit a different level of engagement than those on references for English Wikipedia.

5.2 Average engagement with individual images: image-specific click-through rate

On average, an image in a Wikipedia article gets clicked 2.6 times every 100 impressions. Again, the iCTR is higher (3.2%) for desktop than for mobile users (2.2%). In Fig. 8 we report examples of highly engaging and less engaging images. By visually inspecting these results, we can see some visual trends: highly engaging images seem to depict outdoor environments. In contrast, among the images with low levels of iCTR, we can find human faces.

Figure 8
figure 8

Examples of high and low image-specific CTR images by page popularity (left) and image quality (right). We ranked images by iCTR, popularity and quality, and picked examples from the top-100 (“high”) and bottom-100 (“low”) for each dimension

6 RQ2: what drives reader’s engagement with images when reading Wikipedia articles?

To address RQ2, we now model reader interaction with images on Wikipedia using the factors listed in Sect. 4.3.

6.1 Exploratory analysis

We start our analysis by seeking a relationship between our target metric, the iCTR, and each of the contextual and visual factors in Sect. 4.3. We report the Spearman’s rank correlation coefficients \(\rho _{ctr}\) between the iCTR and the scalar predictors in Fig. 6 and 7. Considering the contextual factors, we observe a negative correlation with article length and popularity (\(\rho =-0.31\) and \(\rho =-0.21\), respectively). When further investigating the relationship with article popularity (Fig. 7), we find that it seems non-linear: engagement with images is low for highly unpopular pages. It becomes higher for pages in a mid-level bucket of popularity and drops again for highly viewed pages. Regarding the image size, despide images are displayed in different resolutions, this does not have a clear relation with the iCTR (\(\rho =-0.002\)). When considering the position in the page instead, the median iCTR is higher for images in galleries (median \(\mathrm{iCTR}=0.024\)) than for images in the infobox (median \(\mathrm{iCTR}=0.019\)) and inline (median \(\mathrm{iCTR}=0.016\)). Moreover, we see signals of reader visual preferences in terms of article topics (Fig. 5(B)): the topics with the highest median value are transportation (0.037) and visual arts (0.037), while politics and sports show the lowest level of interaction with a median iCTR of 0.008. Finally, the correlation analysis of the visual factors confirms our initial intuitions from the visual analysis. There is a positive correlation between the iCTR and outdoor scenery (\(\rho =0.23\)) and a negative relation between the presence of faces and readers’ engagement (\(\rho =-0.14\)). A complete summary of the numerical values discussed can be found in the Additional file 1 (Supplementary Table).

6.2 Regression analysis

Next, we aim to understand how much these features are predictive of reader engagement with images. To do so, we perform a logistic regression analysis that classifies images according to their iCTR.

Study design

We build the training set as follows. We take the median value of iCTR and label the images in our dataset with two classes of high and low iCTR according to whether their iCTR is above or below the median.Footnote 16 We use the contextual and visual factors described in Sect. 4.3 as predictors and the binary iCTR as the target variable. Moreover, we split the predictors into two sets of features and train two separate logistic regression models. The first set of features consists of the topic vectors, while the second consists of the remaining other factors. In the second set of features, we log-transform variables that span over different scales, such as page popularity, text length, caption length, and the number of faces. Moreover, to reduce the amount of multicollinearity among the predictors, we manually inspect the correlation table and compute the Variance Inflation Factor [53] for each variable. We decide to exclude the inline variable, as it shows strong collinearity with gallery and infobox. Finally, we standardize each predictor in the two sets of features.

Impact of image resolution

We found that images on Wikipedia are displayed in different resolutions. Before running the regression analysis, we test the hypothesis that the image size could be decisive in attracting clicks, i. e. readers may tend to click on smaller images as it may be harder to see the details. In Sect. 6.1 we found the correlation coefficient to be −0.002 (with \(p < 0.001\)), indicating no clear relationship between the two variables. Moreover, we observe that image resolution is highly related to its position within the page: the median resolution is about 46, 36, and 11 megapixels respectively for images in the infobox, inline, and in galleries. Also, image resolution is highly correlated with some topics, e. g. it has large positive correlation with biography and entertainment, and large negative correlation with geography and visual arts. Since the image resolution does not seem to be directly related to the iCTR, while it seems to be influenced by some other independent variables, and thus may act as a confounder, we decide not to take it into account in the subsequent analyses.

Controlling for page length and popularity

Similar to what was described in previous work on engagement with Wikipedia content [35], we found that the page popularity and the text length have strong negative correlations with the target. Since page popularity and text length show large variations across the other predictors, especially across topics, we remove the effect of these two confounding variables with a matched study. We build a bipartite graph with images of low and high iCTR as nodes of the two halves. We split the log-transformed page popularity and text length ranges into 100 bins of equal size each, and assign the nodes to these bins, linking two nodes of opposite iCTR when falling into the same bins of popularity and length. Finally, we use min-weight matching on the bipartite graph to find pairs of high/low iCTR samples that minimize the Euclidean distance between all pairs. This procedure succesfully balanced the dataset, with the standardized mean difference of text length and pape popularity across the two classes dropping from −0.54 and −0.51 to −0.010 and −0.007, respectively.

Results

The resulting regression models have an area under the ROC curve (AUC) of 0.67 and 0.62 for the model trained on the topics and the model trained on the other variables, respectively. Figure 9 shows the resulting models coefficients. In Fig. 9(A), we observe that clicks on images are more often related to topics such as transportation, visual arts, geography, and military. On the contrary, clicks on images are less likely in education, sports, and entertainment articles. In Fig. 9(B), we observe that the most important negative predictor is the text offset, i.e. the relative position of the image with respect to the length of the article, meaning that images are more clicked if placed in the upper part of an article. Regarding the visual content, we observe a strong positive effect of outdoor settings, consistently with the positive coefficients of transportation and geography, topics in which a large portion of images display outdoor scenes. Regarding the image position on the page, we find that images in galleries show a high level of engagement, as well as images in the infobox, even though with a moderate effect. Noteworthy, the presence of faces has negative impact in predicting a high level of interactions with images, contrary to what we would expect from the literature [8]. In the remainder of this section, we further investigate this inconsistency in depth, by performing a clustering experiment and an observational study on the images in our dataset.

Figure 9
figure 9

Association of the features with the image iCTR expressed as coefficients of the logistic regressions. (A) Coefficients of the model trained with topics of the article as predictors. (B) Coefficients of the model trained with the other variables of the image. Error bars represent 95% confidence intervals

6.3 Identifying prototypical image groups

To dive deeper into the results emerging from the regression analysis, we provide in this section a non-linear multivariate analysis of our data.

Study design

Our goal is to draw a complementary picture of the complex interplay between reader engagement and image features, identifying prototypical groups of Wikipedia images with homogeneous characteristics. To this extent, we perform a density-based clustering using HDBSCAN [54], which seeks partitions with high density areas of points separated by low density areas, possibly containing noise objects. The advantage of using HDBSCAN is threefold: first, its density-based structure allows to better identify areas of continuous, non-globular points compared to other clustering algorithms that rely on the assumptions of spherical shape clusters, e.g., k-means [55]. Second, by labeling the sparse background points as noise, it aggregates data into coherent clusters rather than partitions. Finally, it extends DBSCAN [56] by implementing a hierarchical clustering approach that allows to extract the optimal flat grouping based on the stability of the clusters, allowing to find groups with non homogeneous density in contrast to a global density threshold adopted by DBSCAN.

We run HDBSCANFootnote 17 on the features set described in Sect. 6.2 including the binary iCTR variable and limiting the analysis to the eight most popular topics (geography, biography, entertainment, visual arts, transportation, sports, military, and biology) that account for 92% of the images in our corpus. HDBSCAN has two main hyper-parameters that have significant practical effect on the clustering: min_cluster_size which refers to the minimum number of grouped items to consider as a cluster, and min_samples which provides a measure of how conservative the clustering would be defining the level at which points are considered noise. The larger the value, the more conservative the clustering, that implies more points will be declared as noise, and clusters will be restricted to progressively more dense areas. We explore the hyper-parameter space with a grid search approach to find the best configuration that maximizes the Density-Based Clustering Validation (DBCV) index [58]. Due to computational constraints, we perform the clustering on a random sample of 50K images, we repeat the procedure 5 times to assess the stability of the tuning phase. We achieve the best configuration with \(\mathit{min}\_\mathit{cluster}\_\mathit{size}=600\) and \(\mathit{min}\_\mathit{samples}=5\) in the majority of the runs. With these settings, we identify 23 clusters, with a number of images ranging between 600 and 5000.

Results

We summarize in Fig. 10 the characteristics of the centroids of the 12 most populated clusters, where each facet represents the mean value of that feature across the examples in that cluster. For ease of visualization, we discretize continuous variables in three classes: low, medium, or high, according to whether the value falls, respectively, in the first, second, or third quantile of the feature distribution. To provide a more clear visual representation of the clusters, we labeled them with descriptive names. We also manually inspected the images in each cluster and chose two to four representative images among the most popular ones. A complete summary of the clustering results can be found in the Additional file 1 (Supplementary Figure).

Figure 10
figure 10

Visual representation of the clustering. The radar plots show for a group centroid the intensity of each feature on a three classes scale. We summarize in green the topics that cover at least 85% of the images categories in a cluster

In the rest of this section, we explore more in depth image quality and its interplay with images containing faces. Even though quality appears, on aggregate, to be moderately positively associated with the tendency to click on images, the underlying phenomenology is more nuanced. On one hand, high-quality images within the geography, transportation, visual arts, military, and biology categories (clusters 2, 3, 5, 6, 7, 8, and 9) show high iCTR across a wide range of contextual factors. A large portion of these images depicts outdoor sceneries that is coherent with the positive coefficient of the outdoor feature in the regression in Sect. 6.2. On the other hand, low quality images are often associated with the presence of faces, especially in topics such as biography, entertainment, and sports, wich overall tend to have a lower click-through rate. Focusing on the interplay between biographies and iCTR reveals significant differences across page popularity and topics worth studying. Images within unpopular biographies, predominantly inline and with a curated textual description, show high iCTR (cluster 10), as well as images placed in galleries in biographies of unpopular artists (cluster 1). On the contrary, popular biographies (cluster 11) or pages that present popular athletes (cluster 12), experience a low iCTR. A possible explanation for this behavior is that users may tend to click on an image in a biography if they do not recognize immediately the subject depicted, while for prominent celebrities, especially if the image is accessible in the infobox, the information need is fulfilled without the need of a click and the interaction with the Media Viewer.

6.4 Are faces engaging on Wikipedia?

As pointed out in Sect. 4.3, images with faces generally elicit high social engagement. In Sect. 6.2, we found that the number of faces has negative weight with respect to the iCTR, while in Sect. 6.3 we observed that Wikipedia readers are more likely to click on images with faces only when placed in less popular biographies. To further investigate this aspect, we design a matched observational study in which we compare the iCTR between images with and without faces. To reduce the effects of confounding factors, we perform a pairwise comparison of images with similar covariates using propensity score matching.

Propensity score matching

Propensity score matching [59] is a statistical technique to evaluate the efficacy of a treatment against a control group, while taking into account the effect of confounding factors. The propensity score is defined as the probability of a sample being treated as a function of the covariates, and it is obtained by training a logistic regression with the covariates as predictors, and the treatment/control variable as target. As a result, observations with the same propensity scores have the same distribution across the observed covariates.

In our experiment, we define images with at least one face as receiving the treatment, images without a face as the control group, and the variables used in the logistic regression (except for the topics and the page popularity) as the covariates.

Results

We consider images in articles about biography, entertainment, and sports, accounting for 90% of all images with at least one face. We find pairs of images minimizing the propensity score within pairs of articles. Figure 11 shows the iCTR as a function of the page popularity, for images with (in orange) and without (in blue) faces. According to a Mann–Whitney U test, the difference between the two distributions is statistically significant, with \(p<0.001\). The tendency to click on images with faces varies depending on page popularity. On pages with less that 1000 monthly pageviews, the presence of faces induces higher level of interactions, with a difference of 0.1%, whereas, after 1000 pageviews, we observe the opposite behavior, with a difference of 0.06%. This also confirms the findings of the clustering analysis.

Figure 11
figure 11

Comparison of the iCTR for images with faces (orange) and without faces (blue) as function of the popularity (pageviews). Error bands represent bootstrapped 95% CIs

To ascertain that our findings remain valid also for non-biographical articles, we replicate the same study by including all the topics in the matching procedure. In this case, we observe a different behavior. Images with faces are less likely to be clicked than others, across all the popularity range. This may explain the overall negative coefficient of the faces feature in the regression analysis, and highlight the role that faces play in increasing engagement on biographical articles.

7 RQ3: do images support reader’s need for additional information when navigating Wikipedia?

We found that readers show a signal of interest in images when reading Wikipedia articles. But are images useful to fulfill part of the reader’s information need when navigating the website? To address this question, we design an additional study that attempts to estimate whether the presence of an image in an article preview can complement the textual information and support in-depth reading.

Matching articles

To check the difference in terms of conversion rate between articles having and not having an image, we first need to reduce the impact of exogenous factors that may potentially drive reader attention on articles, other than the presence of an image. For example, events localized in time can have the effect of sporadically increasing the interest towards specific articles, and therefore on the number of edits [60]. Similarly, the probability of clicking on an article may also depend on its centrality in the article network, i.e. on its in-degree, which is the number of page links pointing to that article. Ideally, we would like to find pairs of articles—one with, the other without image in the preview—that are similar in such factors. To control for these factors, we resort again to propensity score matching. In this experiment, articles with an image in the preview are the treatment group, articles without images are the control, and we use text length, number of edits, and in degree as variables for the matching procedure.

Results

We find pairs of articles by minimizing the propensity score within pairs of articles. Figure 12 shows the conversion rate as a function of article popularity (total number of page views), for articles with (in blue) and without (in yellow) an image in the preview. We find that, according to a Mann–Whitney U test, the difference is statistically significant (\(p<0.001\)), across all the popularity spectrum, with a difference of 2% in the conversion rate.

Figure 12
figure 12

Comparison of the conversion rate for preview tooltip with an image (purple) and without image (green) as function of the page popularity (pageviews). Error bands represent bootstrapped 95% CIs

We rank all pages by conversion rate, and manually inspect the top and bottom articles, with and without images. We find that most of the illustrated articles with higher conversion rate tend to be long lists of aggregated pieces of content related to the same topic, e,g., achievements/publications (movies, books, articles) from notable people or shows. Highly clicked illustrated page previews are often also historical events, or elections, namely information-dense articles where the lead image is only partially useful to grasp the entire article content and its complexity. Conversely, illustrated pages with low conversion rate are articles talking about a specific place (e.g., “Old Fortress, Corfu”), or a specific person, object or spieces (e.g., “Microvelia Macgregori”), namely articles where an illustration can satisfy most of the information need.

Unillustrated page previews with high conversion rate are much more diverse, they go from individual objects or people, e.g. (“Fanny Sidney”), where more textual information is needed to understand the subject in absence an image, to lists and events. Unillustrated articles with lower conversion rate instead tend to be about subjects where a visual explanation is not necessarily needed in order to fully understand the information: for example, generic concept such as “Authority”. “Miniseries”, or “Bachelor of Science”, where images could actually be misleading or give a biased perception of the abstract piece of knowledge.

8 Discussion and conclusions

We provided a comprehensive overview over Wikipedia’s visual world and how readers interact with it. We analyzed reader interactions with visual encyclopedic knowledge and found that images attract more attention than other interactive parts of the article: on average, click-through rate on images is 3.5%, while, for example, reference clicks happen only for 1 in 300 pageviews [35]. Our insights can be summarized as follows:

  • Images serve a cognitive purpose. We found a negative relation between article length and iCTR. This suggests that, similar to references [35], images might be used by readers to complement missing information in the article, fulfilling part of their cognitive function of providing knowledge complementary to the text [2]. Through a matched observational study, we also found that readers tend to click more often on unillustrated Wikipedia page previews to expand their content. On the contrary, conversion rate on illustrated page previews is consistently much lower across popularity buckets, thus suggesting that readers’ need for contextual information is often fulfilled by the presence of an image on the preview popup. In this work, we also tested the relation betwen readers’ interactions with images and article readability: our hypothesis was that images provide a compensatory function for articles that are difficult to read. However, we found evidence of the opposite trend: more readable articles tend to elicit higher engagement with images. While this is a preliminary result, further investigation is needed to understand how images support learning in low readability contexts.

  • We engage more with images illustrating the world and complex objects. Our different layers of analysis consistently expose that Wikipedia readers are attracted by images about geographic locations, especially monuments and maps, and illustrations about biological sciences. Moreover, while we did not explicitly encode the notion of image complexity into our models, we found that Wikipedia readers tend to interact more often with images of complex objects, such as the ones in articles about visual arts, transportation, and military topics. A similar relation between the complexity of the image and its visual interestingness, i.e., the extent to which an image catches the viewer attention, has been widely explored and verified in experimental psychology and computer vision literature [10]. While this relation can be influenced by different visual factors, such as the image size and its content, our results seem to support similar hypothesis, and provide a starting point for further investigation on the relation between image complexity and reader engagement.

  • Faces engage us, but only if unfamiliar. Consistently, research works from different fields suggest that people and web users engage more with faces [61] and face pictures [8], especially celebrities [22], than with other objects or subject, both in online platforms and in the real world. In this work, we found an opposite trend: for Wikipedia readers, images with faces seem to be much less engaging than, for example, more “encyclopedic” images about monuments or transportation. However, we also found that readers do interact with face images when they are placed in unpopular articles, i.e. when those faces represent less well-known people or are unfamiliar. This positive relation between unfamiliarity and engagement again confirms findings from previous research linking the interestingness of a visual object with its familiarity to the observer [10].

Implications

This paper represents a first step towards understanding the importance of images for free knowledge ecosystems. Inspired by theories and ideas from experimental psychology and cognitive science, and by previous studies on Wikipedia readers and web users, our findings describe for the first time how web users interact with the largest source of visual encyclopedic knowledge on the Web. These insights have several implications for different audiences.

For researchers, our results show the feasibility of large scale studies to understand the role of images in instructional settings using a multimodal, computational approach. To this end, experiments could be designed along the same lines of this research, analyzing data coming from, for example, online learning platforms or MOOCs. Researchers could expand the depth and breath of modalities to better understand how and where images should be placed to maximize engagement and learning on the platforms. Researchers could use this work as the basis to build predictive models for image engagement, on Wikipedia and beyond. While our work used basic visual features to understand how readers interact with images, more advanced vision techniques could be used to build end-to-end classifiers that predict the interestingness of an image for Wikipedia readers. This work represents a first step towards understanding the role of images in online instructional settings. While explaining the importance of images in learning is outside the scope of this work, our study shed light on how readers interact with images on Wikipedia, what attracts their attention and which types of visual content they engage with. We look at readers’ interest and usage of images using a fairly implicit, large-scale signal, namely image click-through rate. Future work looking at understanding how readers learn through Wikipedia will need to employ a different set of techniques and signals, i.e., large-scale user studies, focus groups and reading comprehension surveys. This work can be used as a starting point for learning studies. Our feature design is heavily inspired by theories from experimental psychology, computational aesthetics and educational technology research, as well as previous studies analyzing the behavior of Wikipedia readers. Researchers interested in working on learning aspects related to Wikipedia will be able to tap into the same corpus of literature, and look into similar feature design choices.

For editors, given the large amount of unillustrated articles on Wikipedia, and the high level of interest in visual encyclopedic content, the analysis in this paper can help editors prioritize the inclusion of visual content in areas that are highly engaging for Wikipedia readers. Longer term, models and products incoporating signals of readers’ interest in visual content would be extremely helpful for editors. Tools designed to automatically predict reader engagement with images could be incorporated in services and models that help find and prioritize the right images for Wikipedia articles. Given the limited amount of information editors have about how readers interact and learn with Wikipedia content, having visibility over the potential usefuleness of an image in an article would be tremendously helpful to improve editor workflows.

For the broader Wikimedia community, the fact that images help arise interest in free knowledge justifies investments and initiatives designed to improve the pictorial representations of Wikipedia. Our findings on readers interacting with images of monuments and science further encourage the flourishing of initiatives such as Wiki Loves Monuments and Wiki Loves Science which aim at increasing the pictorial representations of these topics. Similarly, the fact that readers are more interested in pictures of unfamiliar people further justifies the existence of organizations such as “Whose Knowledge?”,Footnote 18 who pushes towards the inclusion of visual content in biographies of people from under-represented communities.

For content creators interested in contributing to free knowledge communities and in making their content available in the open, our results provides an initial list of areas of content where closing the visual knowledge gap on Wikipedia [62] is crucial. Knowing that readers tend to be attracted by specific subjects and topics can help the design of new content creation campaigns and donations. The Wikimedia communities, the Wikimedia Foundation, and any web user interested in free knowledge can use these findings to collaborate with GLAM institutions and content creators to make relevant visual content free to use.

Finally, while the scope of this paper is limited to the encyclopedia, Wikipedia represents a central hub of the web ecosystem and the public domain. Its open visual content is re-used across multiple platforms and users, and its images are surfaced at the top of both text and image search results. With this paper, we hope to provide a novel set of results and insights that can build towards better, more open and accessible visual knowledge on Wikipedia, and in turn influence the global accessibility of open visual content in the broader web.

Limitations

While the final goal of our research is to understand images on the broad free knowledge ecosystem, one main limitation of this work is that it mainly focuses on English Wikipedia. With this in mind, we hope in the future to extend this work to include a more representative set of Wikipedia language editions and compare how different language communities interact with visual content.

Most of our analysis depends on the output of existing machine learning models, such as ORES [5], MTCNN [50] or the novel Wikimeda Image Quality classifier. While pretty effective for this task, not all these models have been tested for fairness and inclusivity. As part of our improvements to this work, we would like to employ models that are as debiased as possible and that can be easily applied to images and articles from all around the world.

Readers from different parts of the world come to Wikipedia with different information needs [4]. Additionally, researchers in multimedia computing have shown that different language communities [49] and geographies [63] perceive and produce visual content in different ways. While we focused here on the context and content of Wikipedia images, our analysis completely ignores the characteristics of readers, such as geographic location, internet connection availability for image download, and native language. Our early experiments on global reader behavior show that the way in which readers interact with images on Wikipedia tend to differ across geographic locations, mainly due to broadband availability, modality of access (mobile vs. desktop), and availability of content in their languages. Previous research has indeed shown that the scope and uniqueness of visual material, as well as the availability of content for specific topics largely varies across different language editions [6, 64]. Our next research will extend this analysis to understand, in a privacy-preserving manner, the behavior of different groups of readers with visual encyclopedic content and the impact of exogenous events on image viewership.

Finally, this analysis merely quantifies reader interactions with images, without understanding the actual reason behind the action of clicking on visual content. Our choice of metrics for interest operationalization was driven by an extensive literature studying user interactions with content on web platforms, as reported in Related Work. Click-through rate and conversion rate are widely used to measure image relevance, search satisfaction, user interest in illustrated ads, and reader interactions with citations on Wikipedia. While providing a big picture of readers’ behavior with Wikipedia visual content, a more detailed representation of user interactions could provide complementary insights on this front. Future work will explore a larger set of metrics such as hovers, dwell-time, and eye tracking movements. These metrics are not currently collected by the Wikipedia instrumentation pipeline and we will need to research additional data collection tools. As part of our efforts to understand the importance of images in free knowledge ecosystems, in the future we will also use surveys and user studies to learn why readers look at images on Wikipedia, and further characterize how people use the largest visual encyclopedic knowledge repository.

Availability of data and materials

The user web logs collected during the current study and the quantities computed from them are not publicly available due to privacy restrictions. All the other image features mentioned in the Data Collection section are publicly availailable at the specified urls, or are available from the corresponding author on reasonable request.

Notes

  1. List of Wikipedias. https://meta.wikimedia.org/wiki/List_of_Wikipedias. Accessed March 2021.

  2. Wikimedia Commons. https://commons.wikimedia.org/wiki/Main_Page. Accessed March 2021.

  3. Galleries, Libraries, Archives, and Museums

  4. Image use policy. https://en.wikipedia.org/wiki/Wikipedia:Image_use_policy#Image_content_and_selection. Accessed March 2021.

  5. Manual of Style. https://en.wikipedia.org/wiki/Wikipedia:Manual_of_Style/Images. Accessed March 2021.

  6. The Media Viewer. https://en.wikipedia.org/wiki/Wikipedia:Media_Viewer. Accessed March 2021.

  7. Wiki Loves Monument. https://www.wikilovesmonuments.org/. Accessed March 2021.

  8. Wikipedia Pages Wanting Photos. https://meta.wikimedia.org/wiki/Wikipedia_Pages_Wanting_Photos. Accessed March 2021.

  9. The Webrequest table. https://wikitech.wikimedia.org/wiki/Analytics/Data_Lake/Traffic/Webrequest. Accessed March 2021.

  10. Bot or Not? Identifying “fake” traffic on Wikipedia, Wikimedia Analytics team. https://techblog.wikimedia.org/2020/10/05/bot-or-not-identifying-fake-traffic-on-wikipedia/. Accessed March 2021.

  11. Wikidata topic model. https://github.com/geohci/wikidata-topic-model. Accessed March 2021.

  12. The WikiProject Directory. https://en.wikipedia.org/wiki/Wikipedia:WikiProject_Directory. Accessed March 2021.

  13. Wikipedia:Manual of Style/Images Size. https://en.wikipedia.org/wiki/Wikipedia:Manual_of_Style/Images#Size. Accessed March 2021.

  14. Commons:Quality images. https://commons.wikimedia.org/wiki/Commons:Quality_images. Accessed March 2021.

  15. Commons:Image guidelines. https://commons.wikimedia.org/wiki/Commons:Image_guidelines. Accessed March 2021.

  16. We repeat the logistic regression analysis with different thresholds splitting the two classes, namely we focus on the highest vs. the lowest percentiles of the images according to their iCTR. We find no significant differences on the resulting regression coefficients. Therefore, we choose the median as the cutoff to maximize the presence of images in the analysis.

  17. To run the algorithm, we use the hdbscan Python library [57]: https://hdbscan.readthedocs.io.

  18. Whose Knowledge?. https://whoseknowledge.org/. Accessed March 2021.

References

  1. Anthony D, Smith SW, Williamson T (2009) Reputation and reliability in collective goods: the case of the online encyclopedia Wikipedia. Ration Soc 21(3):283–306

    Article  Google Scholar 

  2. Levie WH, Lentz R (1982) Effects of text illustrations: a review of research. ECTJ 30(4):195–232

    Article  Google Scholar 

  3. Yasseri T, Sumi R, Rung A, Kornai A, Kertész J (2012) Dynamics of conflicts in Wikipedia. PLoS ONE 7(6):38869

    Article  Google Scholar 

  4. Lemmerich F, Sáez-Trumper D, West R, Zia L (2019) Why the world reads Wikipedia: beyond English speakers. In: Proc. International conference on web search and data mining (WSDM)

    Google Scholar 

  5. Halfaker A, Geiger RS (2020) Ores: lowering barriers with participatory machine learning in wikipedia. Proc Human-Computer Interaction (HCI)

  6. He S, Lin AY, Adar E, Hecht BJ (2018) The_tower_of_babel. jpg: diversity of visual encyclopedic knowledge across Wikipedia language editions. In: Proc. International conference on web and social media (ICWSM)

    Google Scholar 

  7. Viegas FB (2007) The visual side of Wikipedia. In: Proc. Hawaii international conference on system sciences (HICSS)

    Google Scholar 

  8. Bakhshi S, Shamma DA, Gilbert E (2014) Faces engage us: photos with faces attract more likes and comments on Instagram. In: Proc. Conference on human factors in computing systems (SIGCHI)

    Google Scholar 

  9. Park JY, O’Hare N, Schifanella R, Jaimes A, Chung C-W (2015) A large-scale study of user image search behavior on the web. In: Proc. Conference on human factors in computing systems (SIGCHI)

    Google Scholar 

  10. Constantin MG, Redi M, Zen G, Ionescu B (2019) Computational understanding of visual interestingness beyond semantics: literature survey and analysis of covariates. ACM Comput Surv 52(2):1–37

    Article  Google Scholar 

  11. Guo D, Zhang S, Wright KL, McTigue EM (2020) Do you get the picture? A meta-analysis of the effect of graphics on reading comprehension. AERA Open 6(1):2332858420901696

    Article  Google Scholar 

  12. Mayer RE (2002) Multimedia learning. In: Psychology of learning and motivation, vol 41, pp 85–139

    Google Scholar 

  13. Khamparia A, Pandey B (2018) Impact of interactive multimedia in e-learning technologies: role of multimedia in e-learning. In: Digital multimedia: concepts, methodologies, tools, and applications, pp 1087–1110

    Chapter  Google Scholar 

  14. Rudolph M (2017) Cognitive theory of multimedia learning. J Online Higher Educ 1(2):1–10

    Google Scholar 

  15. Tempelman-Kluit N (2006) Multimedia learning theories and online instruction. Coll Res Libr 67(4):364–369

    Article  Google Scholar 

  16. Peeck J (1993) Increasing picture effects in learning from illustrated text. Learn Instr 3(3):227–238

    Article  Google Scholar 

  17. Bernard RM (1990) Using extended captions to improve learning from instructional illustrations. Br J Educ Technol 21(3):215–225

    Article  Google Scholar 

  18. Gygli M, Grabner H, Riemenschneider H, Nater F, Van Gool L (2013) The interestingness of images. In: Proc. International conference on computer vision (ICCV)

    Google Scholar 

  19. Khosla A, Das Sarma A, Hamid R (2014) What makes an image popular?. In: Proc. International world wide web conference (WWW)

    Google Scholar 

  20. Ding K, Ma K, Wang S (2019) Intrinsic image popularity assessment. In: Proc. International conference on multimedia (MM)

    Google Scholar 

  21. Zhang W, Wang W, Wang J, Zha H (2018) User-guided hierarchical attention network for multi-modal social image popularity prediction. In: Proc. International world wide web conference (WWW)

    Google Scholar 

  22. Tsikrika T, Diou C (2014) Multi-evidence user group discovery in professional image search. In: Proc. European conference on information retrieval (ECIR)

    Google Scholar 

  23. Jansen BJ (2008) Searching for digital images on the web. J Doc

  24. Huang J, Efthimiadis EN (2009) Analyzing and evaluating query reformulation strategies in web search logs. In: Proc. Conference on information and knowledge management (CIKM)

    Google Scholar 

  25. Jansen BJ, Spink A, Pedersen JO (2004) The effect of specialized multimedia collections on web searching. J Web Eng 3(3–4):182–199

    Google Scholar 

  26. Heald P, Erickson K, Kretschmer M (2015) The valuation of unprotected works: a case study of public domain images on Wikipedia. Harv JL & Tech 29(1):1–31

    Google Scholar 

  27. Erickson K, Perez FR, Perez JR (2018) What is the commons worth? Estimating the value of wikimedia imagery by observing downstream use. In: Proc. International symposium on open collaboration (OpenSym)

    Google Scholar 

  28. Vaidya G, Kontokostas D, Knuth M, Lehmann J, Hellmann S (2015) Dbpedia commons: structured multimedia metadata from the wikimedia commons. In: Proc. International semantic web conference (ISWC)

    Google Scholar 

  29. Ferrada S, Bustos B, Hogan A (2017) Imgpedia: a linked dataset with content-based analysis of wikimedia images. In: Proc. International semantic web conference (ISWC)

    Google Scholar 

  30. Navarrete T, Villaespesa E (2020) Image-based information: paintings in Wikipedia. J Doc

  31. Lehmann J, Müller-Birn C, Laniado D, Lalmas M, Kaltenbrunner A (2014) Reader preferences and behavior on Wikipedia. In: Proc. Conference on hypertext and social media (HT)

    Google Scholar 

  32. Spoerri A (2007) What is popular on Wikipedia and why? First Monday 12(4):1–6

    Google Scholar 

  33. Salutari F, Da Hora D, Dubuc G, Rossi D (2019) A large-scale study of Wikipedia users’ quality of experience. In: Proc. International world wide web conference (WWW)

    Google Scholar 

  34. Singer P, Lemmerich F, West R, Zia L, Wulczyn E, Strohmaier M, Leskovec J (2017) Why we read Wikipedia. In: Proc. International world wide web conference (WWW)

    Google Scholar 

  35. Piccardi T, Redi M, Colavizza G, West R (2020) Quantifying engagement with citations on Wikipedia. In: Proc. The web conference (WWW)

    Google Scholar 

  36. Piccardi T, Redi M, Colavizza G, West R (2021) On the value of Wikipedia as a gateway to the web. In: Proc. The web conference (WWW)

    Google Scholar 

  37. Chapelle O (2014) Modeling delayed feedback in display advertising. In: Proc. Conference on knowledge discovery and data mining (SIGKDD)

    Google Scholar 

  38. Rosales R, Cheng H, Manavoglu E (2012) Post-click conversion modeling and analysis for non-guaranteed delivery display advertising. In: Proc. Conference on web search and data mining (WSDM)

    Google Scholar 

  39. Ta A-P (2015) Factorization machines with follow-the-regularized-leader for ctr prediction in display advertising. In: Proc. International conference on big data (big data)

    Google Scholar 

  40. Richardson M, Dominowska E, Ragno R (2007) Predicting clicks: estimating the click-through rate for new ads. In: Proc. International world wide web conference (WWW)

    Google Scholar 

  41. Edizel B, Mantrach A, Bai X (2017) Deep character-level click-through rate prediction for sponsored search. In: Proc. Conference on research and development in information retrieval (SIGIR)

    Google Scholar 

  42. Lemmerich F, Sáez-Trumper D, West R, Zia L (2019) Why the world reads Wikipedia: beyond English speakers. In: Proc. International conference on web search and data mining (WSDM)

    Google Scholar 

  43. Flesch R (1948) A new readability yardstick. J Appl Psychol 32(3):221

    Article  Google Scholar 

  44. Paranjape A, West R, Zia L, Leskovec J (2016) Improving website hyperlink structure using server logs. In: Proc. Conference on web search and data mining (WSDM)

    Google Scholar 

  45. Schifanella R, Redi M, Aiello LM (2015) An image is worth more than a thousand favorites: surfacing the hidden beauty of Flickr pictures. In: International conference on web and social media (ICWSM)

    Google Scholar 

  46. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proc. Conference on computer vision and pattern recognition (CVPR)

    Google Scholar 

  47. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: Proc. Conference on computer vision and pattern recognition (CVPR)

    Google Scholar 

  48. Huh M, Agrawal P, Efros AA (2016) What makes imagenet good for transfer learning? arXiv preprint 1608.08614

  49. Pappas N, Redi M, Topkara M, Jou B, Liu H, Chen T, Chang S-F (2016) Multilingual visual sentiment concept matching. In: Proc. International conference on multimedia retrieval (ICMR)

    Google Scholar 

  50. Wen Y, Zhang K, Li Z, Qiao Y (2016) A discriminative feature learning approach for deep face recognition. In: Proc. European conference on computer vision (ECCV)

    Google Scholar 

  51. Zagoruyko S, Komodakis N (2016) Wide residual networks. arXiv preprint 1605.07146

  52. Zhou B, Lapedriza A, Khosla A, Oliva A, Torralba A (2017) Places: a 10 million image database for scene recognition. IEEE Trans Pattern Anal Mach Intell 40(6):1452–1464

    Article  Google Scholar 

  53. Kutner MH, Nachtsheim CJ, Neter J, Li W et al. (2005) Applied linear statistical models, vol 5

    Google Scholar 

  54. Campello RJ, Moulavi D, Sander J (2013) Density-based clustering based on hierarchical density estimates. In: Proc. Pacific-Asia conference on knowledge discovery and data mining (PAKDD)

    Google Scholar 

  55. Lloyd S (1982) Least squares quantization in pcm. IEEE Trans Inf Theory 28(2):129–137

    Article  MathSciNet  MATH  Google Scholar 

  56. Ester M, Kriegel H-P, Sander J, Xu X et al. (1996) A density-based algorithm for discovering clusters in large spatial databases with noise. In: International conference on knowledge discovery and data mining (KDD)

    Google Scholar 

  57. McInnes L, Healy J, Astels S (2017) hdbscan: hierarchical density based clustering. J Open Sour Softw 2(11):205

    Article  Google Scholar 

  58. Moulavi D, Jaskowiak PA, Campello RJ, Zimek A, Sander J (2014) Density-based clustering validation. In: Proc. SIAM international conference on data mining (SDM)

    Google Scholar 

  59. Abadie A, Imbens GW (2006) Large sample properties of matching estimators for average treatment effects. Econometrica 74(1):235–267

    Article  MathSciNet  MATH  Google Scholar 

  60. Georgescu M, Kanhabua N, Krause D, Nejdl W, Siersdorfer S (2013) Extracting event-related information from article updates in Wikipedia. In: Proc. European conference on information retrieval (ECIR)

    Google Scholar 

  61. Morton J, Johnson MH (1991) Conspec and conlern: a two-process theory of infant face recognition. Psychol Rev 98(2):164

    Article  Google Scholar 

  62. Redi M, Gerlach M, Johnson I, Morgan J, Zia L (2020) A taxonomy of knowledge gaps for wikimedia projects (second draft). arXiv preprint 2008.12314

  63. You Q, García-García D, Paluri M, Luo J, Joo J (2017) Cultural diffusion and trends in Facebook photographs. In: Proc. International conference on web and social media (ICWSM)

    Google Scholar 

  64. Piccardi T, West R (2021) Crosslingual topic modeling with WikiPDA. In: Proc. The web conference (WWW)

    Google Scholar 

Download references

Acknowledgements

The authors thank the Wikimedia Research Team for their insightful discussions and the Analytics Team for their technical support.

The authors acknowledge the creators of the images included in Fig. 8, Fig. 10, and the Supplementary Figure. Attributions are provided below:

Figure 8:

Detroit Photographic Company (0707) (cropped).jpg, Unknown author, Public domain, via Wikimedia Commons

Ankor Wat temple.jpg, Kheng Vungvuthy, CC BY-SA 4.0, via Wikimedia Commons

Shirdi Sai Baba 3.jpg, Unknown author, Public domain, via Wikimedia Commons

Lana Turner—Marriage is a Private Affair portrait.jpg, MGM, Public domain, via Wikimedia Commons

Denise Richards 2009.1.jpg, Glenn Francis, CC BY-SA 3.0, via Wikimedia Commons

RIAN archive 700096 Pacific fleet vessels’ sortie for combat training.jpg, RIA Novosti archive, CC BY-SA 3.0, via Wikimedia Commons

Alhashimi.jpg, Patrick Makhoul, CC BY-SA 2.0, via Wikimedia Commons

Class-482-NSE-Bank3.jpg, Spsmiler, Public domain, via Wikimedia Commons

2015 Brown SA Goalie Ian Hunter.jpg, Orion 2012, CC0, via Wikimedia Commons

Metaleptus angulatus (38035311795).jpg, Ben Sale, CC BY 2.0, via Wikimedia Commons

Kostomuksha Nature Reserve 2014.jpg, Igor Georgievskiy, CC BY-SA 4.0, via Wikimedia Commons

CominoBlueLagoon.jpg, JarekPT, CC BY-SA 3.0, via Wikimedia Commons

Kjersti Toppe.jpg, Senter partiet, CC BY-SA 2.0, via Wikimedia Commons

Kardinal Kaspar Karel.jpg, Tomas Urban, CC BY-SA 3.0, via Wikimedia Commons

RanellidaeShell.JPG, Toby Hudson, CC BY-SA 3.0, via Wikimedia Commons

Nuestra Senora de Andalucia by Julio Romero de Torres.JPG, Julio Romero de Torres, Public domain, via Wikimedia Commons

SS Eastern Chief (1917).jpg, United States Navy History and Heritage Command photograph, Public domain, via Wikimedia Commons

Giuseppe gene.jpg, Unknown author, Public domain, via Wikimedia Commons

Polscy karykaturzysci.JPG, Unknown author, Public domain, via Wikimedia Commons

Ada English.jpg, Unknown author, CC BY-SA 4.0, via Wikimedia Commons

Figure 10:

01 mucha documentsdecoratifs 1901.jpg, Alphonse Mucha, Public domain, via Wikimedia Commons

Kuniyoshi Utagawa, Mt fuji from Sumida.jpg, Utagawa Kuniyoshi, Public domain, via Wikimedia Commons

La Route tournante en sous-bois, par Paul Cezanne.jpg, Paul Cezanne, Public domain, via Wikimedia Commons

Mykonos City.jpg, Cifo Buscemi, CC0, via Wikimedia Commons

Ladybower Reservoir From Above.jpg, Joel Vardy, CC BY-SA 4.0, via Wikimedia Commons

Bierstadt Albert Old Faithful.jpg, Albert Bierstadt, Public domain, via Wikimedia Commons

Palacio da Alvorada Exterior.JPG, Palacio do Planalto, Attribution, via Wikimedia Commons

Victoria Clock Tower, Liverpool University—geograph.org.uk—374422.jpg, Sue Adair, Liverpool University

View from 555 California Street in San Francisco—panoramio (4).jpg, Eduardo Manchon, CC BY-SA 3.0, via Wikimedia Commons

Jaarbeurs.JPG, Albert kok, CC BY-SA 3.0, via Wikimedia Commons

Apfel-Berlepsch.jpg, Superbass, CC BY-SA 4.0, via Wikimedia Commons

Elements de decor dun immeuble art nouveau (Paris) (4810271270).jpg, Jean-Pierre Dalbera, CC BY 2.0, via Wikimedia Commons

Snettisham HoardDSCF6580.jpg, Johnbod, CC BY-SA 3.0, via Wikimedia Commons

Peasants in an Interior (1661) Adriaen van Ostade.jpg, Adriaen van Ostade, Public domain, via Wikimedia Commons

740 Park Avenue.jpg, Eden, Janine and Jim, CC BY 2.0, via Wikimedia Commons

Arch Titus, Forum Romanum, Rome, Italy.jpg, Jebulon, CC0, via Wikimedia Commons

Elbphilharmonie, Hamburg.jpg, Hackercatxxy, CC BY-SA 4.0, via Wikimedia Commons

Flowers (2425723494) cropped.jpg, Michal Osmenda, CC BY-SA 2.0, via Wikimedia Commons

Coccinella magnifica01.jpg, Gilles San Martin, CC BY-SA 2.0, via Wikimedia Commons

German Shepherd—DSC 0346 (10096362833).jpg, gomagoti, CC BY-SA 2.5, via Wikimedia Commons

BMW 7er (E38) 20090314 front.jpg, M 93, Public domain, via Wikimedia Commons

13-143 JF-17 LBG SIAE 2015 (18984327841).jpg, Eric Salard, CC BY-SA 2.0, via Wikimedia Commons

Delicate arch sunset.jpg, Palacemusic, CC BY-SA 3.0, via Wikimedia Commons

Lake Natron (Tanzania)—2017-03-06 (very early in rainy season)—satellite image (cropped).jpg, Joshua Stevens/NASA, Public domain, via Wikimedia Commons

Arabian Sea map.png, NormanEinstein, Ras67, CC BY-SA 3.0, via Wikimedia Commons

Rob McElhenney and Kaitlin Olson (12063880473).jpg, Sue Lukenbaugh,CC BY-SA 2.0, via Wikimedia Commons

(L-R) Larry Hagman, Ross Perot, Margot Perot and Suzanne Perot at the Rosewood Crescent Club (8392304697).jpg, SMU Central University Libraries, No restrictions, via Wikimedia Commons

Obama family portrait in the Green Room.jpg, Annie Leibovitz/Released by White House Photo Office, Public domain, via Wikimedia Commons

Brad Pitt June 2014 (cropped).jpg, Foreign and Commonwealth Office, CC BY 2.0, via Wikimedia Commons

Elizabeth Olsen1 (cropped).jpg, https://www.flickr.com/photos/tabercil/, CC BY-SA 2.0, via Wikimedia Commons

Argentina celebrando copa (cropped).jpg, Unknown author, Public domain, via Wikimedia Commons

Alonso 2016.jpg, Box Repsol | Flickr, CC BY 2.0, via Wikimedia Commons

Jerami Grant free throw (cropped).jpg, All-Pro Reels, CC BY-SA 2.0, via Wikimedia Commons

Suresh Raina1.jpg, vijay chennupati, CC BY 2.0, via Wikimedia Commons

Supplementary Figure:

B25-1 300.jpg, USAAF, Public domain, via Wikimedia Commons 1990 Nissan 300ZX.jpg, Mike Reyher, CC BY 2.0, via Wikimedia Commons

BoraBora SEtienne.jpg, Samuel Etienne, CC BY-SA 3.0, via Wikimedia Commons

St Helens before 1980 eruption horizon fixed.jpg, Jim Nieland, Public domain, via Wikimedia Commons

Jessica Pare 2014 at Paleyfest.jpg, Dominick D, CC BY-SA 2.0, via Wikimedia Commons

Mae Carol Jemison.jpg, NASA, Public domain, via Wikimedia Commons

William Shu riseconf 2016 (27358004426).jpg, RISE, CC BY 2.0, via Wikimedia Commons

Jordan by Lipofsky 16577.jpg, Steve Lipofsky, CC BY-SA 3.0, via Wikimedia Commons

Hurricanes Quarterback.JPG, BalticHurricanes, CC BY-SA 3.0, via Wikimedia Commons

2019 Final da Copa America 2019—Alisson.jpg, Palacio do Planalto, CC BY 2.0, via Wikimedia Commons

RupertGrint2018.jpg, Sidewalks Entertainment, CC BY 3.0, via Wikimedia Commons

Salt Bae.png, Terron F. Beckham, CC BY 3.0, via Wikimedia Commons

Rose Leslie (March 2013) (headshot).jpg, Suzi Pratt, CC BY-SA 2.0, via Wikimedia Commons

Kit harrington by sachyn mital (cropped 2).jpg, Sachyn, CC BY-SA 3.0, via Wikimedia Commons

Protein HIF1A PDB 1h2k.png, Emw, CC BY-SA 3.0, via Wikimedia Commons

Dehydroepiandrosterone molecule ball.png, Jynto, CC0, via Wikimedia Commons

Iceland-Hatari-ESC2019-002.jpg, Martin Fjellanger, CC BY-SA 4.0, via Wikimedia Commons

Jack Sprat and his wife by Frederick Richardson.jpg, Frederick Richardson, Public domain, via Wikimedia Commons

Incognito Bangkok.jpg, Vairoj Arunyaangkul, CC BY 2.0, via Wikimedia Commons

WarrenGMagnuson (cropped).jpg, Believed to be official senatorial portrait, Public domain, via Wikimedia Commons

1994 Washington senatorial election map.png, JDPEG, CC BY-SA 4.0, via Wikimedia Commons

Ulm-Wiblingen-Fugger.png, OwenBlacker, Public domain, via Wikimedia Commons

Blackwood’s Magazine—1899 cover.jpg, William Blackwood, Public domain, via Wikimedia Commons

I Heard It Through the Grapevine by Marvin Gaye 1968 US single.png, Tamla Records, Public domain, via Wikimedia Commons

AmphibiaLogoTransparent.png, Disney, Public domain, via Wikimedia Commons

SXSW 2016—Rami Malek (25138464364) (cropped).jpg, Daniel Benavides, CC BY 2.0, via Wikimedia Commons

Orlando Bloom Cannes 2013.jpg, Georges Biard, CC BY-SA 3.0, via Wikimedia Commons

Charlize-theron-IMG 6045.jpg, Fuzheado, CC BY-SA 4.0, via Wikimedia Commons

Lorne Michaels David Shankbone 2010.jpg, David Shankbone, CC BY 3.0, via Wikimedia Commons

NYC—Washington Square Park—Arch.jpg, Jean-Christophe Benoist, CC BY 3.0, via Wikimedia Commons

M-S Sarfaq Ittuk.jpg, David Stanley, CC BY 2.0, via Wikimedia Commons

Funding

RS has been partially supported by the project “Countering Online hate speech through Effective on-line Monitoring” funded by the Compagnia di San Paolo. The funder had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

DR, TP, MR, and RS conceptualized the problem and designed research; DR, TP and MR collected data; DR ran the analysis and DR, TP, MR, and RS interpreted the results; DR, TP, MR, and RS wrote the paper. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Daniele Rama.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Supplementary Information

Below is the link to the electronic supplementary material.

13688_2021_312_MOESM1_ESM.pdf

The PDF file entitled “Supplementary Material” contains: a Supplementary Figure for the clustering analysis, depicting all the clusters not in Fig. 10; a Supplementary Table including the numerical values discussed in Sect. 4.3 and 6.1. (PDF 6.8 MB)

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Rama, D., Piccardi, T., Redi, M. et al. A large scale study of reader interactions with images on Wikipedia. EPJ Data Sci. 11, 1 (2022). https://doi.org/10.1140/epjds/s13688-021-00312-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1140/epjds/s13688-021-00312-8

Keywords