Skip to main content
Springer logoLink to Springer
. 2025 Mar 12;243(4):89. doi: 10.1007/s00221-025-07030-x

Elicited emotion: effects of inoculation of an art style on emotionally strong images

Amparo Caceres Gutierrez 1,, Julián Tejada 2, Enrique García Fernández-Abascal 3,4
PMCID: PMC11903571  PMID: 40072572

Abstract

The objective of this research is to study how the application of the Convolutional Neural Network (CNN) artistic filter can be an alternative to mitigate the emotional response to photographs with strong emotional content published in Internet news. Van Gogh’s artistic style was extracted through a CNN and inoculated with 64 IAPS images chosen to cover the entire emotional space. 140 university students of both sexes (70 men and 70 women) with an average age of 22 years, evaluated 128 stimuli, 64 original and 64 digitally inoculated, giving the appearance that they were painted with the artistic style of Van Gogh. For the evaluation of the stimuli, four groups were established under the conditions: 1 high arousal—positive valence, 2 negative valence—low arousal, 3 high arousal—negative valence and 4, low arousal, positive valence. The original images (OI) tended to produce less pleasant effects, while the images inoculated with filters made with CNN provoked reactions with a tendency to calm. The most significant changes in the emotional states are observed in the valence, the stimuli with the inoculation of the artistic style produces alterations with a tendency to pleasant effects. The averages of the comparisons of the dimensions valence and arousal of the OI and the CNN allow to show that there are differences in the emotional states, the results can permit the development of a methodology that, based on the inoculation of the artistic style of original paintings through CNN in emotionally strong images, a new image is created that replaces the strong images published in the Internet news.

Keywords: Artistic filter, Emotional response mitigation, Convolutional neural network

Introduction

The image is a very effective tool to transmit emotions online, the increasing ease of manipulating and distributing images through mobile devices and social networks allows users to share their ideas, thoughts, emotions and experiences with other people generating an emotional response (Lang 1979; Joshi et al., 2011; Jou et al. 2015; You et al. 2015; You et al. 2016; Wang 2020; Shahzad and Alhoori 2021; Wan et al. 2022). Studies show that viewers spend more time on websites with manipulated or fake images. These images, often accompanying online news, can be emotionally distressing and misleading. Because viewers are often unaware that images can be altered, they are more vulnerable to misinformation, which can influence their opinions and actions. (Zillmann et al., 1999; Fahmy et al. 2014; Stumpf, 2017; Kasra et al. 2018; Shen et al. 2019; Knobloch et al. 2003; Vraga et al. 2020; Vieweg 2010; Mohd Shariff et al. 2014; Guidry et al. 2015; Lewandowsky et al. 2012; Lazer et al. 2018; Vázquez- Herrero, 2019).

Visually-focused social media platforms attract long visits, but they also expose users to a higher risk of encountering false information presented through images. (Vieweg 2010; Mohd Shariff et al. 2014; Guidry et al. 2015; Lewandowsky et al. 2012; Lazer et al. 2018; Vázquez-Herrero 2019).

Images in online news serve a dual purpose: to document facts and to evoke emotions. Strong images are particularly effective at stirring emotions, often more so than other forms of media. Therefore, the emotional content is the most important attribute of a news photograph (Sharkey 1993; Graham 2000). Negative images (unpleasant-threatening) have a strong presence in online journalism, the emphasis on visual content, design and aesthetics favor the emotional image, which stimulates user participation, capturing greater attention and impact (Rivers and Mathews 1988; Zillmann et al. 1999; Pavlik 2001; Knobloch et al. 2003; Rössler et al. 2011; Stumpf, 2017; Keib et al. 2018; Li and Xie 2020; Vraga et al. 2020).

The study of the criteria for the selection and impact of press photography (display of strong emotions, broad colours, vivid images) and the effect on the recipients (level of memory, evaluation of the emotional impact of the image and duration of observation) shows the influence on attention to attract and perceive the image (Demarmels 2007; Rössler et al. 2011).

The study on understanding the meaning of photo content shows that photos with faces are more likely to be liked and commented on by Instagram users (Bakhshi et al. 2015). News that contains images attracts more attention and encourages more user engagement than text-only news.,The emotional impact of these images depends on the careful selection of the subject matter and the capturing of a particularly meaningful moment in time,. These are the key elements that make a news photograph powerful (Chiorean 2016; Rivers and Mathews 1988; Keib et al. 2018; Li and Xie 2020).

Fake news spreads faster, further and more often than real news. When selecting journalistic stories, users click more often on certain topics, such as terrorism, natural disasters, or urban legends, specially if these topics contain emotional words classified as trust, anticipation or anger, which can evoke fear, disgust and surprise (Winter and Krämer 2014; Vosoughi et al. 2018; Pröllochs et al. 2021).

In order to detect and respond to the spread of strong photographs in digital news, content moderators and algorithms to monitor and control the dissemination of impactful photographs are used. (Seering et al. 2019; Roberts 2019; Gerrard 2020). Content moderators are people responsible for filtering, flagging, or censoring the content of inappropriate material and potentially dangerous information (Freire et al. 2022). Elements that define the removal criteria include the most extreme end of the graphic content spectrum, such as child exploitation, animal cruelty, suicide and murder. However, despite their best efforts, they cannot prevent all strong images from being filtered (Gillespie et al. 2020).

Algorithmic moderation systems are management mechanisms that structure participation in an online community to facilitate cooperation, prevent abuse, and remove illegal content quickly and effectively (Grimmelmann 2015). The goal of algorithmic moderation is to identify, match, predict, or classify some strong content, such as an image, based on its exact properties or general characteristics. Automated algorithmic moderation tools that platforms use to police content across a wide range of subject areas of scale photography, including terrorism, graphic violence, and toxicity, are applied when scale issues make manual moderation impossible (Gorwa et al. 2020).

Much of the correction of misinformation in online news focuses on the textual content. However, some studies indicate that the use of the Picture-O-Meter barometer allows evaluating whether a photograph is authentic or not, a forensic label is placed in photography to influence the perception of credibility of images (Shen et al. 2019, 2021). To ensure the authenticity of digital images and mitigate misinformation, different watermarking techniques have been developed using algorithms; bit it is difficult to achieve a watermarking system that is both robust and secure (Wan et al. 2022).

Emotion and art style

Emotions are psychophysiological processes that help the organism to adapt to the environment and achieve its goals, they have an adaptive and evolutionary character, since they prepare the organism to emit a behavior in response to a demand from the environment, for example, fear prepares for a fight/flight response, anger for protection or joy for approach (Lang 1995, 2010; Eisenberg 2000). The bioinformational model of emotions defines emotion as a state of predisposition to action based on the activation of one of the two primary motivational systems, appetitive or defensive, which is associated with approach/consummation or escape/avoidance behaviors (Lang 1995).

From this perspective, emotions are dispositions to action that arise in response to stimuli that are meaningful to the individual and are generated in a triple response system: cognitive/experiential/subjective; motor/behavioral/expressive, neurophysiological or cognitive associated with the context in which they develop and the evaluation is carried out from three dimensions, valence, arousal and dominance (Bradley 2009; Lang 1995, 2010; Bradley and Lang 2007).

Emotional experience can be understood through three key dimensions: valence, arousal, and dominance. Valence refers to the positivity or negativity of an emotion, ranging from pleasant to unpleasant. Arousal describes the intensity or activation level of the emotion, ranging from excitement to calmness. Dominance relates to the sense of control a person feels over their emotional response, influencing whether they approach or avoid a situation. (Bradley and Lang 2000; Vila et al. 2001; Coan and Allen 2007).

Emotions and art are intimately linked, from antiquity to the present day, theories of aesthetics have emphasized the role of art in the evocation, configuration and modification of emotions and are at the basis of aesthetic evaluations of objects and works of art (Leder et al. 2004; Silvia 2009; Schindler et al. 2017; Pelowski et al. 2016; Fingerhut and Prinz 2018). Artworks can evoke a wide variety of emotional responses, such as calmness, dynamism, agitation, happiness, anger, fear, disgust, sadness, shock, and erotic desire, these emotions can arise when a person perceives and evaluates a stimulus for its aesthetic appeal (Zhao et al. 2017; Cupchik 2015; Lu et al. 2016).

In painting, humans have mastered the skill of creating unique visual experiences by composing a complex interplay between the content and artistic style of a work of art (Gatys et al. 2015a, b). Artistic style is a multidimensional variable related to the individuality of the artist, the similarity of the aesthetic composition, the space and the harmonization between the particularity of the artist, the trend of the time and society (Jr et al. 2008; Cupchik 2015).

A painter’s unique artistic style is defined by the consistent use of elements like dominant colors, size, line, lighting, texture, shape, and space, as well as the way they utilize their materials. This combination creates the work’s content and shapes the viewer’s aesthetic experience (Zhao et al. 2021). It is common for artists to use the environment as a basis to create works that reflect their emotions and artistic style (Datta et al. 2006; Karayev et al. 2014a, b; Peng 2017; Huang and Belongie 2017; Li et al. 2017; Ionescu et al. 2022; Liao and Huang 2022).

The viewer’s initial encounter with a work of art involves a rapid analysis of its stylistic characteristics. This assessment creates a basic impression of the work’s structure and meaning, stimulating the viewer to interpret its various elements and relate them to their own past experiences in a meaningful way. (Leder et al. 2004; Van Paasschen et al. 2015).

Digital filters and CNN

The use of computer technology to apply digital filters to images posted on the internet has become a way for people to express and communicate affection and emotion, filtered photos on social networks are more likely to be viewed and commented on (Melo and Gratch 2009; Bakhshi et al. 2015). Digital filters are tools that allow you to inoculate color, adjust shadow, texture, light exposure or simulate a change in focus of an image, create special effects such as stylizing, saturating or aging a photo by changing the emotional salience of the image, changing the emotional state of the user (Bakhshi et al. 2015; Chandler and Livingston 2016; Murrin 2017).

The filter significantly changes the look of photography by providing users with convenient and fun ways to distance their images from reality, producing an artistic experience (Trope and Liberman 2010). Studies related to the psychology of the digital filter examine how photographic filters can affect perception and increase viewers’ levels of interpretation by perceiving images in a more abstract and interpretive way (Peng and Iii 2018).

Dark filters, such as black or monotone, are often associated with negative emotions like anger, fear, sadness, and disgust. Conversely, light filters tend to be linked to positive emotions such as surprise, anticipation, trust, and joy. Images with high valence (more positive) are generally more engaging and captivating to viewers, while those with low valence (more negative) can convey feelings of negativity (Murrin 2017).

When predicting the emotional content of images based on valence and arousal, researchers found that certain visual features offer valuable clues. Texture characteristics were strongly associated with calmness. Roundness-angularity and simplicity-complexity were found to correlate equally with both valence and arousal, particularly when comparing images with strong emotional content to neutral images (Lu et al. 2016).

Research on the nostalgic consumption of photography among young people reveals that Instagram consumers give value to their experience of everyday life through the use of digital filters with nostalgic effects applied to contemporary photographs. For the group of participants evaluated, the photographs that look older are perceived as more narrative, because they have more power to tell stories (Morlot 2013).

Studies on the motivation of users to filter photos on the Internet, have shown that they prefer to focus specifically on algorithms that are trendy, popular, aesthetically appealing, aimed at social interaction, or the imagination of the audience, or that control the quality of the photo, correcting errors such as lighting, shadows and sharpness (Hutto et al. 2013; Lavrence and Cambre 2020).

This is how techniques to automate the coding of visual content, through filters,make it possible to understand diffusion patterns and the different characteristics of images to interpret emotional content (Datta et al. 2006; Sermanet et al. 2014; Yuan et al. 2013; Krizhevsky et al. 2014; You et al. 2017; Joo Steinert-Threlkeld 2018; Zhao et al. 2021).

In experimental research, automatic classification algorithms were used to evaluate the aesthetic quality of photographs found on websites. The algorithms extracted features such as light exposure, saturation, and hue. Additionally, the rule of thirds was applied, dividing each image into nine equal parts to identify key points of interest. The results indicated a correlation between these extracted features and the emotions evoked by the images. (Machajdik and Hanbury 2010). Another study used automatic classifiers to discriminate between aesthetically pleasing and unpleasant images, by extracting visual features (familiarity, texture, size and aspect ratio, composition, depth of field, shape convexity) (Datta et al. 2006; Joshi et al. 2011).

Research on how artistic principles influence the emotions conveyed by images has employed various methods. For example, Zhao et al. (2017) used an algorithm to extract artistic style characteristics to classify and evaluate image emotions. In another study, Johnson et al. (2007) analyzed Van Gogh’s post-impressionist style through automatic stroke extraction and color management. Their results demonstrated the expressive effects of his paintings, revealing how color reflected the content, his interactions with other artists, and his emotions.

The aesthetics and emotional values of images are related to their meaning, the extraction of image characteristics such as color, texture and shape allows the recognition of emotions in images and art (Joshi et al. 2011). Likewise, psychological studies show that the aesthetic response of an image depends on dimensions such as composition, color, spatial organization, emphasis, movement or depth (Axelsson 2007; Freeman 2007; Peters 2007).

Studies relating to experimental aesthetics establish that various aspects of color perception play a relevant role in preference judgments, more saturated colors are correlated with positive ratings for valence, images with high valence values will be more interesting, thus captivating the viewer, while images with low values transmit emotions of negativity (Datta et al. 2006; Palmer and Schloss 2010; Mallon et al. 2014).

In this regard, the concept of psychological studies and art theory are used to extract characteristics from images, thus allowing the recognition of a variety of emotions in these images (Machajdik and Hanbury 2010). In the arts, emotionally negative objects can be judged positively, the results relating to emotional experiences (joy, anger, disgust, fear, sadness and shame) and aesthetic judgments show that the aesthetic evaluation of emotionally negative objects can be evaluated in a positive way (Gerger et al. 2014).

Researches related to visual emotions expressed in images establish that through different combinations of aspects of the artistic style of the image such as balance, gradation, movement, rhythm and proportion, emotions can be evoked. symmetrical and harmonious images tend to evoke positive emotions, while images that present strong color contrast can evoke negative emotions (hatred, anguish, anxiety) (Collingwood 1993; Zhao et al. 2017; Ruskan 2021).

Computational methods permit to record in the perceived images, aspects applied to aesthetics and the evocation of emotions. These methods are based on the processing of points or pixels to alter the color and brightness of each pixel (Stork 2009; Peng 2017).One of the methods that stands out are the Convolutional Neural Networks (CNNs), which consists of multiple layers of filters of small computing units that hierarchically process visual information. Each layer of units is understood as a collection of image filters, where each one extracts a certain feature from the input and output image. They are the so-called feature maps, which are versions of the input image with different filters (Nolan 2019).

One of the methods used in the training of the CNN is to decode an original photograph by transferring it to an artistic style of a work of art, the neural algorithm used for the synthesis of the artistic style is an artificial system of neural networks that allow arbitrary separation and recombination the content and style of images, to create high-quality artistic images, controlling the design of the image in an abstract style from the content of a real photograph and the style representation of a work of art (Gatys et al. 2015a, b; Zhao et al. 2018).

This model allows stimulating the production of new results from the transfer of style between images by applying artistic styles of paintings to photographs, (Gatys et al. 2015a) como se observa en la Fig. 1 y Fig. 2.

Fig. 1.

Fig. 1

An example of artistic style transfer. Source images: Tuebingen (Andreas Praefcke, [GFDL or CC BY 3.0], Wikimedia Commons) and Vincent van Gogh’s ‘Starry Night’ (Public Domain). Final stylized image created using neural style transfer, as detailed in the TensorFlow tutorial (https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/tutorials/generative/style_transfer.ipynb#scrollTo=6msVLevwcRhm)

Fig. 2.

Fig. 2

An example of artistic style transfer using different artitis styles: Vassily Kandinsky’s - Composition 7; Gustav Klimt - Thekiss; Piet Mondrian - Composición de colores nº 1 con rojo y azul. Source images: Cat (Adlum Ramadani, [GFDL or CC BY 3.0], Wikimedia Commons). Final stylized image created using neural style transfer, as detailed in the TensorFlow tutorial (https://colab.research.google.com/github/tensorflow/docs/blob/master/site/en/tutorials/generative/style_transfer.ipynb)

Researchers have used CNNs to transfer artistic styles to images, analyzing both the differences and similarities between artistic and non-artistic images, and identifying common properties of visual art (Simon et al. 2014). To capture more complex aspects of the original artistic style, researchers are exploring various approaches. These include modifying how the artistic style is represented, gathering more detailed information about the style, and applying more precise quality constraints to the style transfer results (Novak and Nikulin 2016).

Some studies have used CNN architecture for the analysis of image emotions, for instance, to label the emotional valence of Flickr images, (You et al. 2017); another study used the RMTHG (Rolling Mulli-Task Hypergraph) learning-based method and the combination of algorithms that propose a large-scale classification of Flickr image data for experimentation and testing, personalized emotion perception of images was managed using factors such as visual content, social context, temporal evolution and location (Zhao and Yao, 2016).

At determining the impact of specific regions of a visual image with emotional content, the researchers developed two models from CNN: one oriented to attention mechanisms that allowed an emotion classifier to be built in these regions (You et al. 2017). The other experiment applied CNS (central nervous system) deep metric learning that enabled pattern recognition and analysis for the retrieval and classification of affective images through an emotion vector using texture information (Yang et al. 2018).

The study on photographs and human emotions shows that different people have different emotional reactions to the same image and the person can present multiple emotional reactions to an image, by using CNN that adjust the color tone and texture it is possible to change the evoked emotion of an image (Peng 2017).

It is required to consider and understand human subjectivity and the context in which emotion is perceived, as a result essential relationships are established between computational image analysis and psychology, aesthetics in visual art and photography (Perrett et al. 1994; Zhang et al. 2013; Sartori et al. 2015; Peng 2017; Yang et al. 2018). CNN-based emotion classification methods have demonstrated superior performance in predicting emotions from images posted on the Internet, compared to traditional methods such as natural language processing emotional label classification for images (Cambria et al. 2017; Ahsan et al. 2017; Oliveira et al. 2018; Bonasoli et al. 2018; Xiong et al. 2019; Yadav and Vishwakarma 2020; Alexandru et al. 2022).

Justification

The rapid spread of news with emotionally powerful images on social media, combined with a lack of verification and filtering, makes it difficult for people to identify manipulated images. Viewers often rely on the source, the platform’s interface, their own biases, and accompanying text to judge an image’s authenticity, rather than on the image itself. While some corrections are made, they are infrequent and inconsistent, failing to fully address the widespread misinterpretation of visual information online. (Metzger and Flanagin 2015; Allcott and Gentzkow 2017; Kasra et al. 2018; Meinert et al. 2018; Walter and Tukachinsky 2020).

Neither human moderation nor artificial intelligence are capable of controlling the spread of harmful content from strong images published in Internet news, when toxic online communities on conventional platforms face moderation measures, such as bans, they can migrate to other platforms (Farid 2021; Ribeiro et al. 2021). These aspects raise different questions regarding the emotional effects such as fear, disgust and surprise of strong images published in the news on social networks converted into an environment influenced by emotional distress, distrust, uncertainty, misinformation and deception.

To address the illegal and harmful content of photographs published in social media news, additional technologies are required that filter this content to a more manageable volume that mitigates the psychological effects presented. An alternative is the application of CNN that decode an original photograph, transferring it to an artistic style (Gatys et al. 2015a, b). Studies related to psychology of digital filters examine how photographic filters affect perception and increase viewers’ interpretation levels by perceiving images in a more abstract and interpretive way (Peng and Iii 2018). Two studies (Karunakaran and Ramakrishan 2019; Das et al. 2020) explored how filtering images could lessen their emotional impact on content moderators. They used grayscale and blurring (low-pass) filters and observed a mitigating effect. We anticipate that our more sophisticated “art filter” will provide comparable results. This kind of filters.

significantly changes the look of photography by providing users with convenient and fun ways to distance their images from reality by producing an artistic experience and a new interpretation of the perceived image content (Trope and Liberman 2010). In this context, applying artistic styles to images using Convolutional Neural Networks (CNNs) could offer a way to lessen the emotional impact of powerful visuals on the web. We propose the term “filter inoculation” for this process, suggesting that it might help individuals cope with potentially stressful images or better evaluate the credibility of online information.

Objective

The objective of this research is to study how the application of artistic filters CNN are an alternative to mitigate the emotional response of strong photographs published on the Internet, for which we propose to evaluate the emotional response of images from the IAPS (International Affective Picture System) of the that we already know their emotional assessment in valence or arousal, after having been inoculated with a CNN filter of Van Gogh’s artistic style.

Hypothesis

The perception of emotional valence will change with the application of the CNN art filter, so it is expected that inoculated versions of the IAPS images will be judged as causing less emotional arousal and less unpleasant valence.

Method

The present research aims to use the artificial system based on CNN. It is a filter that captures paintings, identifies their characteristics and apply the artistic style to a photographic image, allowing artistic images of high perceptual quality, consisting of layers of specialized units that are used to process visual information in a hierarchical manner, each of the layers produces different shape maps that are the filtered input image. This method synthesizes the image through the application of Van Gogh’s artistic style, independently captures the content of each photographic image selected from the IAPS, identifies its characteristics and the result is a painting with textures. It is presented by designing new stimuli that introduce two independent perceptually significant sources of variation: the appearance and content of an image (Gatys et al. 2015a, b). His work used a CNN that has the ability to separate the style and content of paintings following the method of stylization proposed by Gatys et al. (2015a, b). As a biologically-inspired model of vision, CNN has demonstrated near-human performance in recognizing objects and faces. Its works examined artworks as a unit constructed by their formal elements of the artwork is regarded as a more defined procedure. This method employs a pre-trained Convolutional Neural Network (CNN), such as VGG (Visual Geometry Group), typically trained on a large image dataset. The network’s hierarchical structure extracts feature maps from different layers, each representing different aspects of the image. A randomly initialized image is then iteratively modified to combine the content of the original image with the desired style, guided by these feature maps. After several iterations, the resulting image retains the original content while adopting the style of the reference artwork. This process, based on the Gatys et al. (2015a, b) method, one of the first CNN-based approaches to image stylization, was selected because it effectively addresses our need to generate stylistic versions of the IAPS images.

We chose Vincent van Gogh’s post-impressionist style for two reasons. Firstly, his paintings, despite their subjective nature, clearly depict physical objects. Secondly, Van Gogh’s prolific body of work provides a large and varied dataset, which is ideal for training the CNN. The Vincent Van Gogh artistic style is characterized by visible, broad, repetitive brush strokes and had the ability to represent objects with a certain level of abstraction. In his paintings, colors were used to capture the mood instead to be used realistically, he often deliberately restricted his palette to a few colors; his paintings include portraits, self-portraits, landscapes and still life (Li et al. 2012). Furthermore, this style is widely used in the CNN field.

Participants

A total of 140 people participated in the study aged between 23 and 25 years, 70 were men and 70 women. This number of participants is compatible with an estimated sample size considering a 2-way ANOVA (2 × 4), a medium effect size and a power of 0.8.

For the experiment, the independent variable (VI) was the emotional stimuli OI and CNN, the dependent variables (DV) were the assessment of the valence and activation dimensions of the OI and the CNN using the 9-point SAM scale. The valence dimension corresponds to the level of pleasant or unpleasant that the image makes you feel; 9 means high level, 5 intermediate point and one lower level. The second dimension corresponded to the activation that goes from the state of alert to the state of calm, 1 means very calm, 5 neutral and 9 very alert.

The independent variables were distributed in four blocks selected under four conditions: Blocks 1 High Activation—High Valence, 2 Low Valence—Low Activation, 3 High Activation—Low Valence and 4 Low Activation, High Valence, following the IAPS affective ratings (lang et al. 2008). The participant results, in terms of valence and activation were compared with the original values of valence and activation reported by lang et al. (2008)

Materials

A total of 64 images were selected from the IAPS bank using their valence and activation as a criterion following the IAPS affective ratings (lang et al. 2008) (see appendix 1) and distributed in four conditions with the same number of images: 1st high activation—high valence (see appendix 2), 2nd low valence—low activation (see appendix 3), 3rd high activation—low valence (see appendix 4) and 4th low activation—high valence (see appendix 5). High-activation images were considered to be those with an average activation between 5.76 and 7.35 (values higher than the third quartile), and those with an average of low activation between 1.72 and 3.97 (values below the first quartile). In terms of valence, mean values between 6.75 and 8.34 (values above the third quartile) were considered high, and mean values between 1.31 and 3.53 (values below the first quartile) were considered low valence. These 64 images were later inoculated into CNN in such a way that each block ended up consisting of 16 OIs and 16 CNN for a total of 128 stimuli. During the filter inoculation process, we discarded any CNN-generated images that significantly deviated from the originals. Images with extraneous details not part of the artistic style were also rejected. This was done to ensure that the context of the original image was preserved.

The OpenSesame 3 software was used, it is an program that allows you to create the experiment, generates graphic experiences, through the tutorial the experiment will be structured step by step. to control the presentation of stimuli and record the participants’ responses, (Mathôt et al. 2012). The task was presented on a SAMSUNG 8 GB DDR4 memory computer, 1 TB 5400 RPM hard drive with a 1366 X 768 resolution display. (Mathôt et al. 2012). The task was presented on a SAMSUNG 8 GB DDR4 memory computer, 1 TB 5400 RPM hard drive with a 1366 X 768 resolution display.

To record the emotional response, the SAM scale (Self-Assessment Maniki: Self-Assessment Scale with Mannequins) (Bradley and Lang 1994; Bradley and Lang 1994) was used, in an iconographic format that represents as directly as possible the emotional reactions of those evaluated in each of the two dimensions (see Fig. 3).

Fig. 3.

Fig. 3

SAM scales of A arousal and B valences used to record participants’ emotional response. Source: (Bradley and Lang 1994)

Design

The experimental task began with the presentation of the instructions on the screen explaining the task and how the SAM valence scales should be used as indicated in Fig. 3 and activation. To do this, the images of each of the SAM scales appeared on the screen with nine numbers aligned at the bottom of the image, indicating to the participant that they had to type a number between 1 and 9 on the keyboard. Low numbers indicate little pleasure/arousal, high numbers indicate a lot of pleasure/arousal, and the number 5 means neutral or intermediate, as indicated in Figs. 4 and 5.

Fig. 4.

Fig. 4

Instructions on how to use the SAM valence scale. Source: (Bradley and Lang 1994)

Fig. 5.

Fig. 5

Screen for instructions on how to use the arousal SAM scale. tags. Source: (Bradley and Lang 1994)

The participants are informed about the development of the experiment related to the evaluation of different emotional images, according to each of the four blocks of images that are randomly presented, each block will have two events: event one corresponds to the presentation of the OI image to be evaluated, event two of each block corresponds to the evaluation of the image inoculated through artistic filters.

Participants sat about 60 cm in front of the monitor and only began the experimental task after having accepted the term of free and informed consent. Once the doubts were clarified, the experimental blocks began, which consisted of the presentation of each set of OIs and their respective CNN version, accompanied by the respective SAM scales. The presentation of each set of images on the monitor begins with a fixation point always followed by the OI and the respective screens for the evaluation of arousal and valence, to subsequently present the CNN image with the respective SAM scales.

We chose to presented the OR version of the image always before the inoculated one, because we want to evaluate how much the level of activation changed when an inoculated version of the same image is presented then immediately. On the hypothesis that the inoculation on the image that already know could cause a change in the emotional response.

The presentation of the images, as well as the scales, had a predetermined time limit (5 s for the OI and their CNN versions and free time to respond to each of the SAM scales), and each valence-combination block arousal occurred randomly. The task ended when participants finished evaluating all 128 images, as shown in Fig. 6.

Fig. 6.

Fig. 6

Diagram of the presentation of each set of OI and CNN images and the respective valence and activation evaluations caused by them.

Source: The author

Data analysis

For data analysis, repeated measurements (2 × 4) were performed with two intra-subject factors: the image factor with two levels, real and CNN, and the block factor with 4 levels, high activation-high valence; low valence and activation; low activation and high valence; Low activation and high valence. The Anova was preceded by normality tests and subsequently post-hoc tests of multiple comparisons were performed using a Bonferroni fit. In all cases, a p-value of less than 0.05 was significant.

Results

The comparison of the results of the assessments was carried out using the experimental design model, in which the blocks, the dimensions (valence and arousal) and the OIs and CNNs are compared, the results allow us to know if the OI represents the same emotional experience compared to the inoculated CNN image. In order to check if there are differences with respect to the valence and arousal dimensions of the OI and the CNNs, ANOVA was used with a significance level of 0.05, testing the hypothesis of equality in the average ratings. The results are as follows: Table 1.

Table 1.

Anova analysis of the valuations of the valence and arousal dimensions of the OI and the CNNs

graphic file with name 221_2025_7030_Tab1_HTML.jpg

The average values between blocks differ Source: The author.

Generally speaking, manipulating images with art-style artificial neural networks decreases valence and arousal. Figure 7 presents a scatter figure that compares the values of valence and arousal attributed to both the OIs and the inoculated images and in which it is possible to observe that the arousal in general decreased, especially for block 4. Something similar occurred with the valence being concentrated around the value of 5, indicating that the inoculated images were perceived as less positive or negative.

Fig. 7.

Fig. 7

Scatter plot comparing the responses to the SAM scales for each of the images.

Source: The autor

The result of the ANOVA shows that there were differences in the way participants perceived the images after inoculation with art-style artificial neural networks, both in terms of valence and arousal.

As can be seen in Table 2, the ANOVA detected significant differences related to the images, the blocks and the image-block interaction. Post-hoc tests indicated that block 4 (low arousal—high valence) was evaluated as causing significantly lower valence than that caused by the images in blocks 1 (high arousal—high valence) and 2 (low valence—low arousal). On the other hand, block 3 (high arousal—low valence) was judged to have a significantly higher valence than blocks 2 and 3 (See Table 2).

Table 2.

Result of the ANOVA model for valence

graphic file with name 221_2025_7030_Tab2_HTML.jpg

P values below 0.05 are considered significant. Source: The author

Regarding valence, significant differences were also found within blocks 1, 2 and 4, meaning that the inoculated images were considered different from those attributed to the original version. In the case of block 1 (high arousal—high valence) the inoculated images were considered to have significantly lower valence, and in the case of blocks 2 (low valence—low arousal) and 4 (low arousal—high valence) the inoculated images were judged to have significantly higher valence (see Fig. 8).

Fig. 8.

Fig. 8

Average valence values for each of the four blocks. Error bars represent the standard error of the mean and horizontal lines represent significant differences at *** = p < 0.001.

Source: The author. Block 1: high arousal-high valence; Block 2: low valence—low arousal; Block 3: high arousal—low valence and Block 4: low arousal—high valence. The horizontal lines represent significant differences between the different blocks/images

Regarding arousal, Table 3 shows that the ANOVA detected significant differences related to the images and the image-block interaction. Post-hoc tests indicated that differences in the perception of arousal caused by the inoculated versions of the images within blocks 2 and 4 were judged by significantly less arousal (see Fig. 9).

Table 3.

Result of the ANOVA activation model.

Source: The autor

graphic file with name 221_2025_7030_Tab3_HTML.jpg

P values less than 0.05 are considered significant

Fig. 9.

Fig. 9

Average activation values for each of the four blocks. Error bars represent the standard error of the mean and horizontal lines represent significant differences at * = p < 0.05, Source: The author. Block 1: high arousal-high valence; Block 2: low valence—low arousal; Block 3: high arousal—low valence and Block 4: low arousal—high valence. The horizontal lines represent significant differences between the different blocks/images

Discussion

The results show that through the evaluations of the OI and the CNNs regarding valence and arousal, no value was zero, which indicates that changes have occurred in the emotional states of the participants. The comparisons between the blocks reflect differences in the averages: block 1 presents the most significant changes in emotional states with more pleasant valence and more arousing pleasant (Bakhshi et al. 2015; Chandler and Livingston 2016; Murrin 2017); block 3 presents trending changes to more pleasantness and less tension (Peng and Iii 2018), with respect to block 2 and 4, similar trends are observed, although with different numerical evaluations in the milder behavioral changes with respect to the valences with less pleasant and less arousing pleasant experiences (Ou et al. 2004).

The averages of the comparisons of the valence and arousal dimensions of the OI and the CNN allow us to show that there are differences in the emotional states, the most significant changes are observed in the valence, where the stimuli produce alterations in the moods with a tendency towards pleasant effects (Murrin 2017); arousal reflects less significant variations with changes in the emotional experience towards calm, the evaluation carried out is a basic aspect of people that allows conclusive distinctions to be made characterizing their experiences in different types of emotions such as pleasure or displeasure (Zhao et al. 2021).

The interactions of the 4 blocks and dimensions valence and activation reflect differences, the most significant are obtained from the interaction between block 1 and the valence similar to block 3 and valence with a tendency towards changes related to pleasant emotional experiences, the interaction between block 4 and activation is similar to the interaction between block 1 and activation reflecting changes towards emotional states with a tendency to calm (Zhao et al. 2017).

The averages in the interactions between blocks and images are different, blocks 1 and 3 reflect representative changes in the evaluations of the OI and the CNN, while the evaluations of blocks 3 and 4 are similar in the OI and the CNN, (Gatys, Ecker and Bethge 2015c). The averages relating to the interactions between the dimensions, the OI and the CNNs reflect differences in the evaluations, the CNN and the valence present the most significant changes, meaning that the effects produced by the CNN are more pleasant, with respect to activation, OIs tend to produce effects of greater displeasure and tension, while CNNs present reactions that tend toward calm. (Davitz 1964; Osgood et al. 1957; Mehrabian and Russell 1974; Joshi et al. 2011).

Conclusion

In summary, the present research demonstrated how the application of artistic filters with CNNs are an alternative to mitigate the emotional response to strong photographs published on the Internet. Using the neural network-based architecture of CNNs they inoculated a OI through artistic filters and participants separately evaluated OIs and inoculated images using the dimensions of emotional valence (unpleasant to pleasant) and arousal (calming to exciting). The ratings show that there were differences in how participants perceived the images; After inoculation with CNN, emotional reactivity decreased, showing that the effects produced by CNNs are more pleasant with respect to the valence and with respect to the arousal of OIs. Uninoculated images tended to produce more tense effects, while images inoculated with filters made with CNN provoked reactions that tended to be calm. The changes in the most significant emotional states are observed in valence, new stimuli produce alterations with a tendency to pleasant effects. Simultaneously, these results may allow the development of another methodology that, based on different artistic filters originating from different artistic styles derived from original paintings, creates a new photographic image that can replace strong photographs published in the news on the Internet. On the other hand, although the changes reflected in arousal present less significant variations, the trend showed a change in the emotional experience towards calm, suggesting that it is possible to create unique experiences through the artistic style of an image.

In this way, the strong photographic image inoculated from the artistic filter with CNN will have an emotional value with less emotional arousal and less negative valence. The results suggest that strong photographs published in Internet news inoculated to a CNN art style can be used to mitigate users emotional response. The findings present a viable way at a theoretical level to relate emotion, photographic image and CNN artistic filters, digital technology applied to media, aesthetics and users, among others.

Our research blends psychology, art, AI, and computer science to investigate how applying artistic styles to images using CNNs can affect viewers’ emotional responses. We call this process “filter inoculation.” The primary goal is to assess the potential of this technique in various fields, such as medicine, communication, and psychotherapy, as a tool for managing reactions to emotionally powerful images. Future research will explore applying and analyzing different artistic styles, potentially mixing them, and investigating alternative inoculation methods. We also plan to broaden the research to include aesthetic emotions and other media, such as music, videos, animations, and films.

Limitations

This study has limitations that should be acknowledged. First, the image set used, drawn from the international affective picture system (IAPS), may not be entirely equivalent to images found in news or on social media. IAPS images often depict naturalistic situations, while online images are frequently more digitally manipulated. However, the IAPS remains one of the most validated sets of images for eliciting emotional responses.

Another limitation is the use of only one artistic style: that of Vincent van Gogh. While this choice aimed for a consistent application of a style that would minimize unexpected emotional reactions due to unfamiliarity, it may have introduced a different bias. Viewers who dislike Van Gogh’s style might experience negative emotions, potentially affecting the study’s results.

Acknowledgement

The authors gratefully acknowledge the National University of Distance Education (UNED) for funding open access publication of this article.

Appendix 1 Image selection criteria, valence and arousal

graphic file with name 221_2025_7030_Figa_HTML.jpg

Appendix 2 Condition 1

graphic file with name 221_2025_7030_Figb_HTML.jpg

Appendix 3 Condition 2

graphic file with name 221_2025_7030_Figc_HTML.jpg

Appendix 4 Condition 3

graphic file with name 221_2025_7030_Figd_HTML.jpg

Appendix 5 Condition 4

graphic file with name 221_2025_7030_Fige_HTML.jpg

Author contributions

Amparo Caceres: wrote the main manuscript text and prepared figures Julian Tejada and Enrique García Fernández-Abascal: made substantial contributions to the conception or design of the work; the acquisition, analysis, or interpretation of data, drafted the work or revised it critically for important intellectual content, approved the version to be published and agree to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Funding

Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature.

Data availability

No datasets were generated or analysed during the current study.

Declarations

Conflict of interest

The authors declare no competing interests.

Footnotes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Ahsan U, De Choudhury M, Essa I (2017) Towards using visual attributes to infer image sentiment of social events. Ga Inst Technol. 10.1109/IJCNN.2017.7966013 [Google Scholar]
  2. Alexandru I, Nicula C, Prodan C, Rotaru RP, Voncilă ML, Tarbă N, Boiangiu CA (2022) Image style transfer via multi-style geometry warping. Appl Sci 12(12):6055. 10.3390/app12126055 [Google Scholar]
  3. Allcott H, Gentzkow M (2017) Social media and fake news in the 2016 election. J Econ Perspectives 31(2):211–236. 10.1257/jep.31.2.211 [Google Scholar]
  4. Axelsson O (2007) Towards a psychology of photography: dimensions underlying aesthetic appeal of photographs. Percept Mot Skills 105:411–434. 10.2466/PMS.105.6.411-434 [DOI] [PubMed] [Google Scholar]
  5. Bakhshi S, Shamma D, Kennedy L, Gilbert E (2015) Why we filter our photos and how it impacts engagement. ICWSM [Google Scholar]
  6. Bonasoli W, Dorini L, Minetto R, Silva TH (2018) Sentiment Analysis in Outdoor Images Using Deep Learning. In: Proceedings of the 24th Brazilian Symposium on Multimedia and the Web. pp. 181–188. 10.1145/3243082.3243093
  7. Bradley MM (2009) Natural selective attention: orienting and emotion. Psychophysiology 46(1):1–11. 10.1111/j.1469-8986.2008.00702.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Bradley MM, Lang PJ (1994) Measuring emotion: the self-assessment manikin and the semantic differential. J Behav Ther Exp Psychiatry 25(1):49–59. 10.1016/0005-7916(94)90063-9 [DOI] [PubMed] [Google Scholar]
  9. Bradley MM, Lang PJ (2000) Measuring emotion: behavior, feeling, and physiology. In: Lane R, Nadel L (eds) Cognitive neuroscience of emotion. Oxford University Press, New York, pp 242–276 [Google Scholar]
  10. Bradley MM, Lang PJ (2007) Emotion and motivation. In: Cacioppo J, Tassinary L, Berntson G (eds) The Handbook of Psychophysiology. Cambridge University Press, New York [Google Scholar]
  11. Cambria E, Poria S, Gelbukh A, Thelwall M (2017) Sentiment analysis is a big suitcase. IEEE Intell Syst 32(6):74–80. 10.1109/MIS.2017.4531228 [Google Scholar]
  12. Chandler L, Livingston D (2016) Reframing the authentic: photography, mobile technologies and the visual language of digital imperfection. Learning to see: The meanings, modes and methods of visual literacy. Brill. 10.1163/9781848883024_019 [Google Scholar]
  13. Chiorean C (2016) T. A. L. A. S. M. A. N. (s. f.). The emotion of a press picture awarded the Pulitzer. Recuperado 8 de julio de 2022, de https://www.academia.edu/27117080/The_emotion_of_a_press_picture_awarded_the_Pulitzer
  14. Coan JA, Allen JJB (2007) Handbook of emotion elicitation and assessment. Oxford University Press [Google Scholar]
  15. Collingwood RG (1993) Los Principios Del Arte. Fondo de Cultura Económica [Google Scholar]
  16. Cupchik G (2015) Theoretical foundations for an empirical aesthetics. pp. 60–85. 10.1017/CBO9781139207058.005
  17. Das A, Dang B, Lease M (2020) Fast, accurate, and healthier: interactive blurring helps moderators reduce exposure to harmful content. 10.26153/TSW/10199
  18. Datta R, Joshi D, Li J, Wang JZ (2006) Studying aesthetics in photographic images using a computational approach. In: Leonardis A, Bischof H, Pinz A (eds) Computer Vision – ECCV 2006. Springer, Berlin, pp 288–301. 10.1007/11744078_23 [Google Scholar]
  19. Davitz JR (1964) The communication of emotional meaning. Mcgraw Hill, p 214 [Google Scholar]
  20. Demarmels S (2007) Review: Katrin Döveling (2005). Emotionen – Medien – Gemeinschaft. Eine kommunikationssoziologische Analyse [Emotions—Media—Society. A Communication-Sociological Analysis]. Forum Qualitative Sozialforschung / Forum: Qualitative Social Research, 8(3), Art. 3. 10.17169/fqs-8.3.295
  21. Eisenberg N (2000) Emotion, regulation, and moral development. Annu Rev Psychol 51(1):665–697. 10.1146/annurev.psych.51.1.665 [DOI] [PubMed] [Google Scholar]
  22. Fahmy S, Bock M, Wanta W (2014) Visual Communication Theory and Research: A Mass Communication Perspective. Palgrave Macmillan US, New York [Google Scholar]
  23. Farid H (2021) An overview of perceptual hashing. J Online Trust Safety. 10.54501/jots.v1i1.24 [Google Scholar]
  24. Fingerhut J, Prinz J (2018) Grounding evaluative concepts. Philos Trans R Soc b: Biol Sci 373:20170142. 10.1098/rstb.2017.0142 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Freeman M (2007) The photographer’s eye: composition and design for better digital photos. CRC Press, Routledge [Google Scholar]
  26. Freire GLS, Panoutsos T, Perez L, Benevenuto F, Figueiredo F (2022) Understanding Effects of Moderation and Migration on Online Video Sharing Platforms. In: Proceedings of the 33rd ACM Conference on Hypertext and Social Media, pp. 220–224. 10.1145/3511095.3536377
  27. Gatys LA, Ecker AS, Bethge M (2015b) Texture synthesis using convolutional neural networks. Adv Neural Inf Process Syst. 10.48550/arXiv.1505.07376 [Google Scholar]
  28. Gatys LA, Ecker AS, Bethge M (2015) A Neural Algorithm of Artistic Style. arXiv:1508.06576 [cs, q-bio]. http://arxiv.org/abs/1508.06576
  29. Gerger G, Leder H, Kremer A (2014) Context effects on emotional and aesthetic evaluations of artworks and IAPS pictures. Acta Physiol (Oxf) 151:174–183. 10.1016/j.actpsy.2014.06.008 [DOI] [PubMed] [Google Scholar]
  30. Gerrard Y (2020) Behind the screen: content moderation in the shadows of social media. New Media Soc 22(3):579–582. 10.1177/1461444819878844 [Google Scholar]
  31. Gillespie T, Aufderheide P, Carmi E, Gerrard Y, Gorwa R, Matamoros-Fernández A, Roberts ST, Sinnreich A, West SM (2020) Expanding the debate about content moderation: Scholarly research agendas for the coming policy debates. Internet Policy Review, 9(4). https://policyreview.info/articles/analysis/expanding-debate-about-content-moderation-scholarly-research-agendas-coming-policy
  32. Gorwa R, Binns R, Katzenbach C (2020) Algorithmic content moderation: technical and political challenges in the automation of platform governance. Big Data Soc 7(1):2053951719897945. 10.1177/2053951719897945 [Google Scholar]
  33. Graham ME (2000) Information Seeking Behaviour in Image Retrieval: Visor 1 Final Report. Imprint unknown.
  34. Grimmelmann J (2015) The Virtues of Moderation. Cornell Law Faculty Publications. https://scholarship.law.cornell.edu/facpub/1486
  35. Guidry JPD, Carlyle K, Messner M, Jin Y (2015) On pins and needles: how vaccines are portrayed on Pinterest. Vaccine 33(39):5051–5056. 10.1016/j.vaccine.2015.08.064 [DOI] [PubMed] [Google Scholar]
  36. Huang X, Belongie S (2017) Arbitrary style transfer in real-time with adaptive instance normalization. arXiv. 10.48550/arXiv.1703.06868 [Google Scholar]
  37. Hutto CJ, Yardi S, Gilbert E (2013) A longitudinal study of follow predictors on twitter. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, 821–830. 10.1145/2470654.2470771
  38. Ionescu B, Bainbridge WA, Murray N (2022) Human Perception of Visual Information: Psychological and Computational Perspectives. Springer Nature [Google Scholar]
  39. Johnson J, Hendriks E, Berezhnoy I, Brevdo E, Hughes S, Daubechies I, Li J, Postma E, Wang J (2008) Image processing for artist identification. Signal Process Mag, IEEE 25:37–48. 10.1109/MSP.2008.923513 [Google Scholar]
  40. Johnson CR, Jr Hendriks E, Berezhnoy JI, Brevdo E, Hughes SM, Daubechies I, Li J, Posma E, Wang JZ (2007) Image Processing for Artist Identification. http://digitalpaintinganalysis.org/literature/ProceedingsIA4AI-1.pdf
  41. Joo J, Steinert-Threlkeld ZC (2018) Image as data: automated visual content analysis for political science. arXiv. 10.48550/arXiv.1810.01544 [Google Scholar]
  42. Joshi D, Datta R, Fedorovskaya E, Luong QT, Wang JZ, Li J, Luo J (2011) Aesthetics and emotions in images. IEEE Audio and Electroacoustics Newsletter 28(5):94–115. 10.1109/MSP.2011.941851 [Google Scholar]
  43. Jou B, Chen T, Pappas N, Redi M, Topkara M, Chang S-F (2015) Visual Affect Around the World: A Large-scale Multilingual Visual Sentiment Ontology. In: Proceedings of the 23rd ACM international conference on Multimedia, 159–168. 10.1145/2733373.2806246
  44. Karayev S, Trentacoste M, Han H, Agarwala A, Darrell T, Hertzmann A, Winnemoeller H (2014) (s. f.). Recognizing Image Style: Extended Abstract. 1
  45. Karayev S, Trentacoste M, Han H, Agarwala A, Darrell T, Hertzmann A, Winnemoeller H (2014) Recognizing Image Style. In: Proceedings of the British Machine Vision Conference 2014, 122.1–122.11. 10.5244/C.28.122
  46. Karunakaran S, Ramakrishan R (2019) Testing stylistic interventions to reduce emotional impact of content moderation workers. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing 7:50–58. 10.1609/hcomp.v7i1.5270 [Google Scholar]
  47. Kasra M, Shen C, O’Brien JF (2018) Seeing Is Believing: How People Fail to Identify Fake Images on the Web. Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, 1–6. 10.1145/3170427.3188604
  48. Keib K, Espina C, Lee Y-I, Wojdynski BW, Choi D, Bang H (2018) Picture this: the influence of emotionally valenced images, on attention, selection, and sharing of social media news. Media Psychol 21(2):202–221. 10.1080/15213269.2017.1378108 [Google Scholar]
  49. Knobloch S, Hastall M, Zillmann D, Callison C (2003) Imagery effects on the selective reading of internet newsmagazines. Commun Res 30(1):3–29. 10.1177/0093650202239023 [Google Scholar]
  50. Krizhevsky A, Sutskever I, Hinton GE (2014) ImageNet Classification with Deep Convolutional Neural Networks. In: Advances in Neural Information Processing Systems, 25. https://papers.nips.cc/paper/2012/hash/c399862d3b9d6b76c8436e924a68c45b-Abstract.html
  51. Lang PJ (1979) A bio-informational theory of emotional imagery. Psychophysiology 16(6):495–512. 10.1111/j.1469-8986.1979.tb01511.x [DOI] [PubMed] [Google Scholar]
  52. Lang PJ (1995) The emotion probe: studies of motivation and attention. Am Psychol 50(5):372–385 [DOI] [PubMed] [Google Scholar]
  53. Lang PJ (2010) Emotion and motivation: toward consensus definitions and a common research purpose. Emot Rev 2(3):229–233. 10.1177/1754073910361984 [Google Scholar]
  54. Lang PJ, Bradley MM, Cuthbert BN (2008) International affective picture system (IAPS): Affective ratings of pictures and instruction manual. Technical Report A-8, Gainesville, FL: University of Florida
  55. Lavrence, C., Cambre, C. (2020). “Do I Look Like My Selfie?” Filters and the Digital-Forensic Gaze. Social Media + Society, 6(4), 2056305120955182. 10.1177/2056305120955182
  56. Lazer D, Baum M, Benkler Y, Berinsky A, Greenhill K, Menczer F, Metzger M, Nyhan B, Pennycook G, Rothschild D, Schudson M, Sloman S, Sunstein C, Thorson E, Watts D, Zittrain J (2018) The science of fake news. Science 359:1094–1096. 10.1126/science.aao2998 [DOI] [PubMed] [Google Scholar]
  57. Leder H, Belke B, Oeberst A, Augustin D (2004) A model of aesthetic appreciation and aesthetic judgments. Br J Psychol (London, England: 1953) 95(Pt 4):489–508. 10.1348/0007126042369811 [DOI] [PubMed] [Google Scholar]
  58. Lewandowsky S, Ecker UKH, Seifert CM, Schwarz N, Cook J (2012) Misinformation and its correction: continued influence and successful debiasing. Psychol Sci Public Interest 13(3):106–131. 10.1177/1529100612451018 [DOI] [PubMed] [Google Scholar]
  59. Li Y, Xie Y (2020) Is a picture worth a thousand words? An empirical study of image content and social media engagement. J Mark Res 57(1):1–19. 10.1177/0022243719881113 [Google Scholar]
  60. Li J, Yao L, Hendriks E, Wang JZ (2012) Rhythmic brushstrokes distinguish van Gogh from his contemporaries: findings via automated brushstroke extraction. IEEE Trans Pattern Anal Mach Intell 34(6):1159–1176. 10.1109/TPAMI.2011.203 [DOI] [PubMed] [Google Scholar]
  61. Li Y, Fang C, Yang J, Wang Z, Lu X, Yang M-H (2017) Universal style transfer via feature transforms. arXiv. 10.48550/arXiv.1705.08086 [Google Scholar]
  62. Liao Y-S, Huang C-R (2022) Semantic context-aware image style transfer. IEEE Trans Image Process 31:1911–1923. 10.1109/TIP.2022.3149237 [DOI] [PubMed] [Google Scholar]
  63. Lu X, Sawant N, Newman MG, Adams RB, Wang JZ, Li J (2016) Identifying Emotions Aroused from Paintings. In: Hua G, Jégou H (eds) Computer Vision – ECCV 2016 Workshops. Springer International Publishing, pp 48–63. 10.1007/978-3-319-46604-0_4 [Google Scholar]
  64. Machajdik J, Hanbury A (2010) Affective image classification using features inspired by psychology and art theory. In: Proceedings of the International Conference on Multimedia - MM ‘10, 83. 10.1145/1873951.1873965
  65. Mallon B, Redies C, Hayn-Leichsenring G (2014) Beauty in abstract paintings: perceptual contrast and statistical properties. Front Human Neurosci. 10.3389/fnhum.2014.00161 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Mathôt S, Schreij D, Theeuwes J (2012) OpenSesame: an open-source, graphical experiment builder for the social sciences. Behav Res Methods 44(2):314–324. 10.3758/s13428-011-0168-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Mehrabian A, Russell JA (1974) An approach to environmental psychology. The MIT Press, pp xii, 266 [Google Scholar]
  68. Meinert J, Aker A, Krämer N (2018) The Impact of Twitter Features on Credibility Ratings-An Explorative Examination Combining Psychological Measurements and Feature Based Selection Methods.
  69. Melo C, Gratch J (2009) The effect of color on expression of joy and sadness in virtual humans. pp. 1–7. 10.1109/ACII.2009.5349585
  70. Metzger MJ, Flanagin AJ (2015) Psychological approaches to credibility assessment online. Handbook Psychol Commun Technol. 10.1002/9781118426456.CH20 [Google Scholar]
  71. 10.1007/978-3-319-06028-6_50
  72. Morlot E (2013) Nostalgic consumption behaviours among young generations in photography: A comparative approach of Instagram and analogue photography. Undefined. https://www.semanticscholar.org/paper/Nostalgic-consumption-behaviours-among-young-in-%3A-A-Morlot/91fae8cb312e49a430706059916fad2a020d85ed
  73. Murrin D (2017) The effect of photo filters on the emotional resonance of teenagers. South Carolina Junior Academy of Science. https://scholarexchange.furman.edu/scjas/2017/all/250
  74. Nolan F (2019) A Neural Algorithm of Artistic Style: A Modern Form of Creation. Medium. https://towardsdatascience.com/a-neural-algorithm-of-artistic-style-a-modern-form-of-creation-d39a6ac7e715
  75. Novak R, Nikulin Y (2016) Improving the Neural Algorithm of Artistic Style. arXiv:1605.04603 [cs]. http://arxiv.org/abs/1605.04603
  76. Oliveira W, Dorini L, Minetto R, Silva T (2018) Sentiment Analysis in Outdoor Images Using Deep Learning. pp. 181–188. 10.5753/webmedia.2018.4574
  77. Osgood CE, Suci GJ, Tannenbaum PH (1957) The measurement of meaning. Univer. Illinois Press, p 342 [Google Scholar]
  78. Ou LC, Luo MR, Woodcock A, Wright A (2004) A study of colour emotion and colour preference. Part 1: colour emotions for single colours. Color Res Appl 29(3):232–240. 10.1002/col.20010 [Google Scholar]
  79. Palmer SE, Schloss KB (2010) An ecological valence theory of human color preference. Proc Natl Acad Sci 107(19):8877–8882. 10.1073/pnas.0906172107 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Pavlik J (2001) Journalism and new media. Columbia University Press. 10.7312/pavl11482 [Google Scholar]
  81. Pelowski M, Markey PS, Lauring JO, Leder H (2016) Visualizing the impact of art: an update and comparison of current psychological models of art experience. Front Hum Neurosci 10:160. 10.3389/fnhum.2016.00160 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Peng Y, Iii JBJ (2018) Feast for the eyes: effects of food perceptions and computer vision features on food photo popularity. Int J Commun 12:24 [Google Scholar]
  83. Peng Y (2017) Time Travel with One Click: Effects of Digital Filters on Perceptions of Photographs. 10.1145/3025453.3025810
  84. Perrett DI, May KA, Yoshikawa S (1994) Facial shape and judgements of female attractiveness. Nature 368(6468):239–242. 10.1038/368239a0 [DOI] [PubMed] [Google Scholar]
  85. Peters G (2007) Aesthetic Primitives of Images for Visualization. In: 2007 11th International Conference Information Visualization (IV ‘07). 10.1109/IV.2007.20
  86. Pröllochs N, Bär D, Feuerriegel S (2021) Emotions explain differences in the diffusion of true vs. False social media rumors. Sci Rep. 10.1038/s41598-021-01813-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Ribeiro MH, Jhaver S, Zannettou S, Blackburn J, De Cristofaro E, Stringhini G, West R (2021) Do Platform Migrations Compromise Content Moderation? Evidence from r/The_Donald and r/Incels. In: Proceedings of the ACM on Human-Computer Interaction, 5(CSCW2). pp. 1-24. 10.1145/34760
  88. Rivers WL, Mathews C (1988) Ethics for the media. Prentice Hall. Commercial Content Moderation: Digital Laborers’ Dirty Work. Media Studies Publications. https://ir.lib.uwo.ca/commpub/12
  89. Roberts ST (2019) Behind the screen: content moderation in the shadows of social media. Yale University Press. 10.2307/j.ctvhrcz0v [Google Scholar]
  90. Rössler P, Bomhoff J, Haschke JF, Kersten J, Müller R (2011) Selection and impact of press photography. An empirical study on the basis of photo news factors. Communications. 10.1515/comm.2011.021 [Google Scholar]
  91. Ruskan J (2021) EMOTION and ART: Mastering The Artist’s Path
  92. Sartori A, Culibrk, Peng K-C, Chen T, Sadovnik A, Gallagher A (2015) A mixed bag of emotions: Model, predict, and transfer emotion distributions. pp. 860–868. 10.1109/CVPR.2015.7298687
  93. Schindler I, Hosoya G, Menninghaus W, Beermann U, Wagner V, Eid M, Scherer KR (2017) Measuring aesthetic emotions: a review of the literature and a new assessment tool. PLoS ONE. 10.1371/journal.pone.0178899 [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Seering J, Wang T, Yoon J, Kaufman G (2019) Moderator engagement and community development in the age of algorithms. New Media Soc 21:146144481882131. 10.1177/1461444818821316 [Google Scholar]
  95. Sermanet P, Eigen D, Zhang X, Mathieu M, Fergus R, LeCun Y (2014) OverFeat: integrated recognition localization and detection using convolutional networks. arXiv. 10.48550/arXiv.1312.6229 [Google Scholar]
  96. Shahzad M, Alhoori H (2021) Public reaction to scientific research via twitter sentiment prediction. J Data Inform Sci. 10.2478/jdis-2022-0003 [Google Scholar]
  97. Sharkey J (1993) When pictures drive foreign policy. Am Journal Rev 15(10):14–20 [Google Scholar]
  98. Shen C, Kasra M, Pan W, Bassett GA, Malloch Y, O’Brien JF (2019) Fake images: the effects of source, intermediary, and digital media literacy on contextual assessment of image credibility online. New Media Soc 21(2):438–463. 10.1177/1461444818799526 [Google Scholar]
  99. Shen C, Kasra M, O’Brien JF (2021) Research note: this photograph has been inoculated: testing the effectiveness of image forensic labeling on news image credibility. Harvard Kennedy School Misinform Rev. 10.37016/mr-2020-72 [Google Scholar]
  100. Silvia PJ (2009) Looking past pleasure: anger, confusion, disgust, pride, surprise, and other unusual aesthetic emotions. Psychol Aesthet Creat Arts 3(1):48–51. 10.1037/a0014632 [Google Scholar]
  101. Simon M, Rodner E, Denzler J (2014) Part Detector Discovery in Deep Convolutional Neural Networks. arXiv:1411.3159 [cs]. http://arxiv.org/abs/1411.3159
  102. Stork DG (2009) Computer vision and computer graphics analysis of paintings and drawings: an introduction to the literature. In: Proceedings of the 13th International Conference on Computer Analysis of Images and Patterns. pp. 9–24. 10.1007/978-3-642-03767-2_2
  103. Stumpf SF-T, Andrew MacFarlane, Ayşe Göker, Simone (2017) Selecting and tailoring of images for visual impact in online journalism [Text]. University of Borås. http://informationr.net/ir/22-1/colis/colis1619.html
  104. Trope Y, Liberman N (2010) Construal-level theory of psychological distance. Psychol Rev 117(2):440–463. 10.1037/a0018963 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Van Paasschen J, Bacci F, Melcher D (2015) The influence of art expertise and training on emotion and preference ratings for representational and abstract artworks. PLoS ONE 10:e0134241. 10.1371/journal.pone.0134241 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Vázquez-Herrero J (2019) Redes, filtros y algoritmos: aproximación al fotoperiodismo en Instagram. Razón y Palabra 23(106):272–289 [Google Scholar]
  107. Vieweg S (2010) Microblogged Contributions to the Emergency Arena: Discovery, Interpretation and Implications
  108. Vila J, Sánchez M, Ramírez I, Fernández M, Cobos P, Rodríguez S, Moltó J (2001) El sistema internacional de imágenes afectivas (IAPS): adaptación española. Segunda parte. Rev Psicol Gen Apl 54(4):635–657 [Google Scholar]
  109. Vosoughi S, Roy D, Aral S (2018) The spread of true and false news online. Science (New York, N.y.) 359(6380):1146–1151. 10.1126/science.aap9559 [DOI] [PubMed] [Google Scholar]
  110. Vraga EK, Kim SC, Cook J, Bode L (2020) Testing the effectiveness of correction placement and type on instagram. Harvard Int J Press/politics 25(4):632–652. 10.1177/1940161220919082 [Google Scholar]
  111. Walter N, Tukachinsky R (2020) A meta-analytic examination of the continued influence of misinformation in the face of correction: how powerful is it, why does it happen, and how to stop it? Commun Res 47(2):155–177. 10.1177/0093650219854600 [Google Scholar]
  112. Wan W, Wang J, Zhang Y, Li J, Yu H, Sun J (2022) A comprehensive survey on robust image watermarking. Neurocomputing 488:226–247. 10.1016/j.neucom.2022.02.083 [Google Scholar]
  113. Wang JZ (2020) Modeling aesthetics and emotions in visual content: from vincent van gogh to robotics and vision. In: Joint Workshop on Aesthetic and Technical Quality Assessment of Multimedia and Media Analytics for Societal Trends. pp. 15–16. 10.1145/3423268.3423586
  114. Winter S, Krämer NC (2014) A question of credibility—effects of source cues and recommendations on information selection on news sites and blogs. Communications 39(4):435–456. 10.1515/commun-2014-0020 [Google Scholar]
  115. Xiong H, Liu Q, Song S, Cai Y (2019) Region-based convolutional neural network using group sparse regularization for image sentiment classification. EURASIP J Image Video Process 2019(1):30. 10.1186/s13640-019-0433-8 [Google Scholar]
  116. Yadav A, Vishwakarma DK (2020) Sentiment analysis using deep learning architectures: a review. Artif Intell Rev 53(6):4335–4385. 10.1007/s10462-019-09794-5 [Google Scholar]
  117. Yang J, She D, Sun M, Cheng M-M, Rosin PL, Wang L (2018) Visual sentiment prediction based on automatic discovery of affective regions. IEEE Trans Multimedia 20(9):2513–2525. 10.1109/TMM.2018.2803520 [Google Scholar]
  118. You Q, Luo J, Jin H, Yang J (2016) Building a large scale dataset for image emotion recognition: the fine print and the benchmark. arXiv. 10.48550/arXiv.1605.02677 [Google Scholar]
  119. You Q, Luo J, Jin H, Yang J (2015) Robust Image Sentiment Analysis Using Progressively Trained and Domain Transferred Deep Networks. arXiv: 1509.06041 [cs]. http://arxiv.org/abs/1509.06041
  120. You Q, Jin H, Luo J (2017) Visual Sentiment Analysis by Attending on Local Image Regions. In: Proceedings of the AAAI Conference on Artificial Intelligence, 31(1), Art. 1. 10.1609/aaai.v31i1.10501
  121. Yuan J, Mcdonough S, You Q, Luo J (2013) Sentribute: Image sentiment analysis from a mid-level perspective. In: Proceedings of the Second International Workshop on Issues of Sentiment Discovery and Opinion Mining, 1–8. 10.1145/2502069.2502079
  122. Zhang H, Yang Z, Gönen M, Koskela M, Laaksonen J, Honkela T, Oja E (2013) Affective Abstract Image Classification and Retrieval Using Multiple Kernel Learning. 10.1007/978-3-642-42051-1_22
  123. Zhao S, Yao H, Gao Y, Ding G, Huang Q, Chua T-S (2016) Predicting personalized image emotion perceptions in social networks. IEEE Transact Affect Comput. 10.1145/2964284.2964289 [Google Scholar]
  124. Zhao S, Yao H, Gao Y, Ji R, Ding G (2017) Continuous probability distribution prediction of image emotions via multitask shared sparse regression. IEEE Trans Multimed 19(3):632–645. 10.1109/TMM.2016.2617741 [Google Scholar]
  125. Zhao S, Ding G, Huang Q, Chua T-S, Schuller B, Keutzer K (2018) Affective image content analysis: a comprehensive survey. IJCAI. 10.24963/ijcai.2018/780 [Google Scholar]
  126. Zhao S, Yao X, Yang J, Jia G, Ding G, Chua T-S, Schuller BW, Keutzer K (2021) Affective image content analysis: two decades review and new perspectives. IEEE Transact Pattern Anal Mach Intell. 10.48550/arXiv.2106.16125 [DOI] [PubMed] [Google Scholar]
  127. Zillmann D, Gibson R, Sargent SL (1999) Effects of photographs in news-magazine reports on issue perception. Media Psychol 1(3):207–228. 10.1207/s1532785xmep0103_2 [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

No datasets were generated or analysed during the current study.


Articles from Experimental Brain Research are provided here courtesy of Springer

RESOURCES