Visual Sentiment Analysis from Disaster Images in Social Media
Abstract
:1. Introduction
- We extend the concept of visual sentiment analysis to a more challenging and crucial task of disaster analysis, generally involving multiple objects and other relevant information in the background of images, and propose a deep architecture-based visual sentiment analyzer for an automatic sentiment analysis of natural disaster-related images from social media.
- Assuming that the available deep architectures respond differently to an image by extracting diverse but complementary image features, we evaluate the performance of several deep architectures pre-trained on ImageNet and Places dataset both individually and in combination.
- We conduct a crowd-sourcing study to analyze people’s sentiments towards disasters and disaster-related content and annotate the training data. In the study, a total of 2338 users participated in analyzing and annotating 4003 disaster-related images (All images are Creative Commons licensed).
- We provide a benchmark visual sentiment analysis dataset with four different sets of annotations, each aimed at solving a separate task, which is expected to be proven as a useful resource for future work in the domain. To the best of our knowledge, this is the first attempt on the subject.
2. Motivation, Concepts, Challenges and Applications
- Defining/identifying sentiments—The biggest challenge in this domain is defining sentiments and identifying the one that better suits the given visual content. Sentiments are very subjective and vary from person to person. Moreover, the intensity of the sentiments conveyed by an image is another item to be tackled.
- Semantic gap—One of the open questions that researchers have thoroughly investigated in the past decades is the semantic gap between visual features and cognition [13]. The selection of visual features is very crucial in multimedia analysis in general and in sentiment analysis in particular. We believe object and scene-level features could help in extracting such visual cues.
- Data collection and annotation—Image sources, sentiment labels, and feature selection are application-dependent. For example, an entertainment or education context is completely different from the humanitarian one. Such diversity makes it difficult to collect benchmark datasets from which knowledge can be transferred, thus requiring ad-hoc data crawling and annotation.
3. Related Work
4. Proposed Visual Sentiment Analysis Processing Pipeline
4.1. Data Collection and Sentiment Category Selection
4.2. The Crowd-Sourcing Study
4.3. Deep Visual Sentiment Analyzer
5. Experiments and Evaluations
5.1. Statistics of the Crowd-Sourcing Study and Dataset
5.2. Datasets
5.3. Experimental Setup
5.4. Experimental Results
5.5. Lessons Learned
- Sentiment analysis aims to extract people’s perceptions of the images; thus, crowd-sourcing seems a suitable option for collecting training and ground truth datasets. However, choosing labels/tags for conducting a successful crowd-sourcing study is not straightforward.
- The most commonly used three tags, namely positive, negative, and neutral, are not enough to fully exploit the potential of visual sentiment analysis in applications such as disaster analysis. The complexity of the task increases as we go deeper into the sentiment/emotion hierarchy.
- The majority of the disaster-related images in social media represent negative (i.e., sad, horror, pain, anger, and fear) sentiments; however, we noticed that there exists a number of samples able to evoke positive emotions, such as joy and relief.
- Disaster-related images from social media exhibit sufficient features to evoke human emotions. The objects in images (gadgets, clothes, broken houses, scene-level (i.e., background, landmarks)), color/contrast, and human expressions, gestures, and poses provide crucial cues in the visual sentiment analysis of disaster-related images. This can be a valuable aspect to be considered to represent people’s emotions and sentiments.
- Human emotions and sentiment tags are correlated, as can also be noticed from the statistics of the crowd-sourcing study. Thus, a multi-label framework is likely to be the most promising research direction.
6. Conclusions and Future Research Directions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Öztürk, N.; Ayvaz, S. Sentiment analysis on Twitter: A text mining approach to the Syrian refugee crisis. Telemat. Inform. 2018, 35, 136–147. [Google Scholar] [CrossRef]
- Kušen, E.; Strembeck, M. Politics, sentiments, and misinformation: An analysis of the Twitter discussion on the 2016 Austrian presidential elections. Online Soc. Netw. Media 2018, 5, 37–50. [Google Scholar] [CrossRef]
- Sadr, H.; Pedram, M.M.; Teshnehlab, M. A Robust Sentiment Analysis Method Based on Sequential Combination of Convolutional and Recursive Neural Networks. Neural Process. Lett. 2019, 50, 2745–2761. [Google Scholar] [CrossRef]
- Barrett, L.F.; Adolphs, R.; Marsella, S.; Martinez, A.M.; Pollak, S.D. Emotional expressions reconsidered: Challenges to inferring emotion from human facial movements. Psychol. Sci. Public Interest 2019, 20, 1–68. [Google Scholar] [CrossRef] [Green Version]
- Poria, S.; Majumder, N.; Hazarika, D.; Cambria, E.; Gelbukh, A.; Hussain, A. Multimodal Sentiment Analysis: Addressing Key Issues and Setting Up the Baselines. IEEE Intell. Syst. 2018, 33, 17–25. [Google Scholar] [CrossRef] [Green Version]
- Said, N.; Ahmad, K.; Riegler, M.; Pogorelov, K.; Hassan, L.; Ahmad, N.; Conci, N. Natural disasters detection in social media and satellite imagery: A survey. Multimed. Tools Appl. 2019, 78, 31267–31302. [Google Scholar] [CrossRef] [Green Version]
- Imran, M.; Ofli, F.; Caragea, D.; Torralba, A. Using AI and Social Media Multimodal Content for Disaster Response and Management: Opportunities, Challenges, and Future Directions. Inf. Process. Manag. 2020, 57, 102261. [Google Scholar] [CrossRef]
- Ahmad, K.; Pogorelov, K.; Riegler, M.; Conci, N.; Halvorsen, P. Social media and satellites. Multimed. Tools Appl. 2018, 78, 2837–2875. [Google Scholar] [CrossRef]
- Hassan, S.Z.; Ahmad, K.; Al-Fuqaha, A.; Conci, N. Sentiment analysis from images of natural disasters. In Proceedings of the International Conference on Image Analysis and Processing, Trento, Italy, 9–13 September 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 104–113. [Google Scholar]
- Chua, T.H.H.; Chang, L. Follow me and like my beautiful selfies: Singapore teenage girls’ engagement in self-presentation and peer comparison on social media. Comput. Hum. Behav. 2016, 55, 190–197. [Google Scholar] [CrossRef]
- Munezero, M.D.; Montero, C.S.; Sutinen, E.; Pajunen, J. Are they different? Affect, feeling, emotion, sentiment, and opinion detection in text. IEEE Trans. Affect. Comput. 2014, 5, 101–111. [Google Scholar] [CrossRef]
- Kim, H.R.; Kim, Y.S.; Kim, S.J.; Lee, I.K. Building emotional machines: Recognizing image emotions through deep neural networks. IEEE Trans. Multimed. 2018, 20, 2980–2992. [Google Scholar] [CrossRef] [Green Version]
- Soleymani, M.; Garcia, D.; Jou, B.; Schuller, B.; Chang, S.F.; Pantic, M. A survey of multimodal sentiment analysis. Image Vis. Comput. 2017, 65, 3–14. [Google Scholar] [CrossRef]
- Khan, K.; Khan, R.U.; Ahmad, K.; Ali, F.; Kwak, K.S. Face Segmentation: A Journey From Classical to Deep Learning Paradigm, Approaches, Trends, and Directions. IEEE Access 2020, 8, 58683–58699. [Google Scholar] [CrossRef]
- Badjatiya, P.; Gupta, S.; Gupta, M.; Varma, V. Deep Learning for Hate Speech Detection in Tweets. In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 April 2017; pp. 759–760. [Google Scholar] [CrossRef] [Green Version]
- Araque, O.; Gatti, L.; Staiano, J.; Guerini, M. DepecheMood++: A Bilingual Emotion Lexicon Built through Simple Yet Powerful Techniques. IEEE Trans. Affect. Comput. 2019. [Google Scholar] [CrossRef] [Green Version]
- Zhang, L.; Wang, S.; Liu, B. Deep learning for sentiment analysis: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1253. [Google Scholar] [CrossRef] [Green Version]
- Ortis, A.; Farinella, G.; Battiato, S. Survey on Visual Sentiment Analysis. IET Image Process. 2020, 14, 1440–1456. [Google Scholar] [CrossRef]
- Machajdik, J.; Hanbury, A. Affective image classification using features inspired by psychology and art theory. In Proceedings of the 18th ACM International Conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 83–92. [Google Scholar]
- Borth, D.; Ji, R.; Chen, T.; Breuel, T.; Chang, S.F. Large-Scale Visual Sentiment Ontology and Detectors Using Adjective Noun Pairs. In Proceedings of the 21st ACM International Conference on Multimedia, Barcelona, Spain, 21–25 October 2013; Association for Computing Machinery: New York, NY, USA, 2013; pp. 223–232. [Google Scholar] [CrossRef] [Green Version]
- Chen, T.; Borth, D.; Darrell, T.; Chang, S.F. Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. arXiv 2014, arXiv:1410.8586. [Google Scholar]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Chandrasekaran, G.; Hemanth, D.J. Efficient Visual Sentiment Prediction Approaches Using Deep Learning Models. In Proceedings of the Iberoamerican Knowledge Graphs and Semantic Web Conference, Kingsville, TX, USA, 22–24 November 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 260–272. [Google Scholar]
- Pournaras, A.; Gkalelis, N.; Galanopoulos, D.; Mezaris, V. Exploiting Out-of-Domain Datasets and Visual Representations for Image Sentiment Classification. In Proceedings of the 2021 16th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP), Corfu, Greece, 4–5 November 2021; pp. 1–6. [Google Scholar]
- Al-Halah, Z.; Aitken, A.P.; Shi, W.; Caballero, J. Smile, Be Happy :) Emoji Embedding for Visual Sentiment Analysis. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 4491–4500. [Google Scholar]
- Huang, F.; Wei, K.; Weng, J.; Li, Z. Attention-Based Modality-Gated Networks for Image-Text Sentiment Analysis. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 16, 1–19. [Google Scholar] [CrossRef]
- Gelli, F.; Uricchio, T.; He, X.; Del Bimbo, A.; Chua, T.S. Learning subjective attributes of images from auxiliary sources. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2263–2271. [Google Scholar]
- You, Q.; Jin, H.; Luo, J. Visual sentiment analysis by attending on local image regions. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
- Ou, H.; Qing, C.; Xu, X.; Jin, J. Multi-Level Context Pyramid Network for Visual Sentiment Analysis. Sensors 2021, 21, 2136. [Google Scholar] [CrossRef]
- Wu, L.; Zhang, H.; Deng, S.; Shi, G.; Liu, X. Discovering Sentimental Interaction via Graph Convolutional Network for Visual Sentiment Prediction. Appl. Sci. 2021, 11, 1404. [Google Scholar] [CrossRef]
- Yadav, A.; Vishwakarma, D.K. A deep learning architecture of RA-DLNet for visual sentiment analysis. Multimed. Syst. 2020, 26, 431–451. [Google Scholar] [CrossRef]
- Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
- Wang, X.; Jia, J.; Yin, J.; Cai, L. Interpretable aesthetic features for affective image classification. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 3230–3234. [Google Scholar]
- Ortis, A.; Farinella, G.M.; Torrisi, G.; Battiato, S. Exploiting objective text description of images for visual sentiment analysis. Multimed. Tools Appl. 2021, 80, 22323–22346. [Google Scholar] [CrossRef]
- Katsurai, M.; Satoh, S. Image sentiment analysis using latent correlations among visual, textual, and sentiment views. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 2837–2841. [Google Scholar]
- Wang, J.; Fu, J.; Xu, Y.; Mei, T. Beyond Object Recognition: Visual Sentiment Analysis with Deep Coupled Adjective and Noun Neural Networks. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI), New York, NY, USA, 9–15 July 2016; pp. 3484–3490. [Google Scholar]
- Peng, K.C.; Sadovnik, A.; Gallagher, A.; Chen, T. Where do emotions come from? Predicting the emotion stimuli map. In Proceedings of the 2016 IEEE international conference on image processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 614–618. [Google Scholar]
- Cowen, A.S.; Keltner, D. Self-report captures 27 distinct categories of emotion bridged by continuous gradients. Proc. Natl. Acad. Sci. USA 2017, 114, E7900–E7909. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zhou, B.; Lapedriza, A.; Xiao, J.; Torralba, A.; Oliva, A. Learning deep features for scene recognition using places database. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 487–495. [Google Scholar]
- Ahmad, K.; Conci, N. How Deep Features Have Improved Event Recognition in Multimedia: A Survey. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2019, 15, 39. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. Densenet: Implementing efficient convnet descriptor pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
- Tan, M.; Le, Q.V. Efficientnet: Rethinking model scaling for convolutional neural networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
- Lemaître, G.; Nogueira, F.; Aridas, C.K. Imbalanced-learn: A Python Toolbox to Tackle the Curse of Imbalanced Datasets in Machine Learning. J. Mach. Learn. Res. 2017, 18, 1–5. [Google Scholar]
Refs. | Dataset | Application | Features/Model | Main Focus |
---|---|---|---|---|
[19] | CMUMOSEI [19] | Generic | Color and texture features, such as Tamura | Relies on features based on psychology and art theory to classify the emotional response of a given image. The features are grouped by color, texture, composition, and content, and then classified by a naive Bayes-based classifier. |
[25] | SimleyNet [25] | Emojis | CNNs | Mainly focuses on detection and classification of emojis, which act as a proxy for the emotional response of an image. Moreover, also proposes a dataset containing over 4 million images and emoticon pairs from Twitter. |
[31] | Twitter dataset [32], Flickr and Instagram dataset [33], CMUMOSEI [19] | Generic | CNNs | Proposes a residual
attention-based deep learning network (RA-DLNet) aiming to learn the spatial hierarchies of image features extracted through CNNs. Moreover, analyses of the performance of seven state-of-the-art CNN architectures on several datasets. |
[34] | Flickr dataset [35] | Generic | CNNs, handcrafted features (GIST, BoW) | Jointly utilize text (objective text description of images obtained/extracted from the visual content of images through CNNs model) and visual features in an embedding space obtained with Canonical Correlation Analysis (CCA). |
[26] | Self-collected dataset | Generic | CNNs, LSTM | Proposes an attention-based network, namely Attention-based Modality-Gated Networks (AMGN) to exploit the correlation between visual and textual information for sentiment analysis. |
[28] | VSO [20] | Generic | CNNs | Aims to explore the role of local image regions in visual sentiment analysis through an attention mechanism -based model. |
[29] | Twitter [36], Flickr and Instagram dataset [33], Emotion ROI [37] | Generic | CNNs | Proposes a multi-level context pyramid network (MCPNet) aiming to combine local and global features in cross-layer feature fusion scheme for better representation of visual cues. |
[30] | Emotion ROI [37], Twitter [36] | Generic | CNNs and GCN | Proposes a Graph Convolutional Network (GCN)- based framework to incorporate the interaction features among different objects in an image. The visual features of the objects represent nodes, while the emotional distances between objects correspond to the edges of the graph. |
This Work | Self-collected | Natural Disasters | CNNs | Explores a new application of visual sentiment analysis by collecting and sharing a benchmark dataset with baseline experimental results. Moreover, it also highlights the key research challenges, potential applications, and stack-holders. |
Sets | Tags |
---|---|
Set 1 | Positive, Negative, Neutral |
Set 2 | Relax, Stimulated, Normal |
Set 3 | Joy, Sadness, Fear, Disgust, Anger, Surprise, and Neutral |
Set 4 | Anger, Anxiety, craving, Empathetic pain, Fear, Horror, Joy, Relief, Sadness, and Surprise |
Tags | # Samples |
---|---|
Positive | 803 |
Negative | 2297 |
Neutral | 579 |
Tags | # Samples | Tags | # Samples |
---|---|---|---|
Joy | 1207 | Sadness | 3336 |
Fear | 2797 | Disgust | 1428 |
Anger | 1419 | Surprise | 2233 |
Neutral | 1892 | - | - |
Tags | # Samples | Tags | # Samples |
---|---|---|---|
Anger | 2108 | Anxiety | 2716 |
Craving | 1100 | Pain | 2544 |
Fear | 2803 | Horror | 2042 |
Joy | 1181 | Relief | 1356 |
Sadness | 3300 | Surprise | 1975 |
Model | Accuracy | Precision | Recall | F-Score |
---|---|---|---|---|
VGGNet (ImageNet) | 92.12 | 88.64 | 87.63 | 87.89 |
VGGNet (Places) | 92.88 | 89.92 | 88.43 | 89.07 |
Inception-v3 (ImageNet) | 82.59 | 76.38 | 68.81 | 71.60 |
ResNet-50 (ImageNet) | 90.61 | 86.32 | 85.18 | 85.63 |
ResNet-101 (ImageNet) | 90.90 | 86.79 | 85.84 | 86.01 |
DenseNet (ImageNet) | 85.77 | 79.39 | 78.53 | 78.20 |
EfficientNet (ImageNet) | 91.31 | 87.00 | 86.94 | 86.70 |
VGGNet (places + ImageNet) | 92.83 | 89.67 | 88.65 | 88.97 |
Model | Accuracy | Precision | Recall | F-Score |
---|---|---|---|---|
VGGNet (ImageNet) | 82.61 | 84.12 | 80.28 | 81.66 |
VGGNet (Places) | 82.94 | 82.87 | 82.30 | 82.28 |
Inception-v3 (ImageNet) | 80.67 | 80.98 | 82.98 | 80.72 |
ResNet-50 (ImageNet) | 82.48 | 84.33 | 79.41 | 81.38 |
ResNet-101 (ImageNet) | 82.70 | 82.92 | 82.04 | 82.20 |
DenseNet (ImageNet) | 81.99 | 83.43 | 81.30 | 81.51 |
EfficientNet (ImageNet) | 82.08 | 82.80 | 81.31 | 81.51 |
VGGNet (places + ImageNet) | 83.18 | 83.13 | 83.04 | 82.57 |
Model | Accuracy | Precision | Recall | F-Score |
---|---|---|---|---|
VGGNet (ImageNet) | 82.74 | 80.43 | 85.61 | 82.14 |
VGGNet (Places) | 81.55 | 79.26 | 85.08 | 81.16 |
Inception-v3 (ImageNet) | 81.53 | 78.21 | 89.30 | 82.27 |
ResNet-50 (ImageNet) | 82.30 | 79.90 | 84.18 | 81.60 |
ResNet-101 (ImageNet) | 82.56 | 80.25 | 84.51 | 81.80 |
DenseNet (ImageNet) | 81.72 | 79.40 | 85.35 | 81.63 |
EfficientNet (ImageNet) | 82.25 | 80.83 | 82.70 | 81.39 |
VGGNet (places + ImageNet) | 82.08 | 79.36 | 87.25 | 81.99 |
Model | Metric | Negative | Neutral | Positive |
---|---|---|---|---|
VGGNet | Accuracy | 88.61 | 95.36 | 91.66 |
Precision | 88.45 | 93.20 | 84.56 | |
Recall | 74.59 | 93.29 | 91.83 | |
F1-Score | 80.85 | 93.22 | 88.04 | |
VGGNet (p) | Accuracy | 90.07 | 94.88 | 93.21 |
Precision | 88.63 | 91.13 | 89.87 | |
Recall | 79.52 | 94.21 | 89.88 | |
F1-Score | 83.79 | 92.64 | 89.85 | |
Inception V-3 | Accuracy | 76.48 | 86.51 | 82.28 |
Precision | 70.64 | 79.34 | 78.25 | |
Recall | 45.76 | 82.51 | 66.86 | |
F1-Score | 55.46 | 80.85 | 71.41 | |
ResNet-50 | Accuracy | 86.95 | 92.22 | 92.07 |
Precision | 83.40 | 87.15 | 88.14 | |
Recall | 74.51 | 90.68 | 88.29 | |
F1-Score | 78.65 | 88.86 | 88.170 | |
ResNet-101 | Accuracy | 87.16 | 92.31 | 92.29 |
Precision | 86.57 | 86.07 | 87.99 | |
Recall | 71.38 | 92.80 | 89.15 | |
F1-Score | 78.11 | 89.25 | 88.54 | |
DenseNet | Accuracy | 80.59 | 87.84 | 87.72 |
Precision | 76.98 | 80.33 | 83.04 | |
Recall | 60.16 | 87.01 | 79.54 | |
F1-Score | 66.15 | 83.10 | 81.18 | |
EfficientNet | Accuracy | 87.50 | 93.91 | 91.66 |
Precision | 86.41 | 93.91 | 84.87 | |
Recall | 72.87 | 92.58 | 91.68 | |
F1-Score | 78.96 | 91.24 | 88.07 | |
VGGNet (P+I) | Accuracy | 89.94 | 94.90 | 92.99 |
Precision | 88.99 | 90.62 | 89.62 | |
Recall | 78.44 | 95.17 | 89.58 | |
F1-Score | 83.15 | 92.81 | 89.58 |
Model | Metric | Joy | Sadness | Fear | Diguest | Anger | Surprise | Neutral |
---|---|---|---|---|---|---|---|---|
VGGNet | Accuracy | 83.37 | 95.32 | 88.24 | 76.67 | 76.86 | 75.29 | 75.78 |
Precision | 92.17 | 92.46 | 85.09 | 76.78 | 82.13 | 76.96 | 80.31 | |
Recall | 76.78 | 99.12 | 94.83 | 60.63 | 56.71 | 77.22 | 73.32 | |
F1-Score | 83.77 | 95.67 | 89.68 | 67.68 | 66.99 | 77.07 | 76.35 | |
VGGNet (p) | Accuracy | 84.59 | 95.67 | 88.86 | 76.07 | 77.43 | 75.99 | 77.21 |
Precision | 92.44 | 93.47 | 85.47 | 71.19 | 76.23 | 75.21 | 81.58 | |
Recall | 78.65 | 98.60 | 95.46 | 67.15 | 62.18 | 82.27 | 75.64 | |
F1-Score | 84.99 | 95.97 | 90.19 | 68.78 | 68.33 | 78.58 | 78.43 | |
Inception V-3 | Accuracy | 79.81 | 90.51 | 85.40 | 76.26 | 75.51 | 76.21 | 75.51 |
Precision | 89.81 | 86.77 | 81.19 | 86.36 | 86.12 | 71.72 | 75.66 | |
Recall | 72.09 | 96.53 | 94.88 | 49.30 | 49.87 | 92.06 | 80.48 | |
F1-Score | 79.94 | 91.39 | 87.50 | 62.57 | 62.64 | 80.62 | 77.84 | |
ResNet-50 | Accuracy | 85.59 | 95.03 | 87.97 | 75.16 | 77.64 | 73.75 | 75.72 |
Precision | 94.16 | 92.71 | 86.18 | 73.83 | 81.89 | 79.31 | 78.49 | |
Recall | 79.15 | 98.19 | 92.52 | 61.23 | 59.70 | 69.63 | 75.59 | |
F1-Score | 85.99 | 95.37 | 89.22 | 66.43 | 68.83 | 73.92 | 76.91 | |
ResNet-101 | Accuracy | 85.10 | 95.30 | 88.38 | 76.13 | 76.10 | 75.43 | 77.24 |
Precision | 88.15 | 93.59 | 86.84 | 76.67 | 76.90 | 74.76 | 79.28 | |
Recall | 84.86 | 97.67 | 92.42 | 58.95 | 61.13 | 82.00 | 77.96 | |
F1-Score | 86.42 | 95.59 | 89.54 | 66.49 | 67.94 | 78.17 | 78.60 | |
DenseNet | Accuracy | 83.81 | 93.51 | 87.32 | 76.24 | 76.48 | 75.78 | 75.43 |
Precision | 91.41 | 91.79 | 85.50 | 81.24 | 87.74 | 73.15 | 77.47 | |
Recall | 78.47 | 96.16 | 92.07 | 53.72 | 50.52 | 86.83 | 76.85 | |
F1-Score | 84.41 | 93.92 | 88.66 | 64.52 | 64.02 | 79.38 | 76.94 | |
EfficientNet | Accuracy | 84.40 | 94.84 | 88.38 | 75.70 | 75.78 | 74.83 | 75.67 |
Precision | 91.44 | 93.04 | 86.16 | 75.49 | 78.24 | 74.24 | 80.47 | |
Recall | 79.73 | 97.41 | 93.52 | 63.65 | 59.57 | 81.50 | 72.67 | |
F1-Score | 85.09 | 95.16 | 89.63 | 67.57 | 66.82 | 77.65 | 76.13 | |
VGGNet (P+I) | Accuracy | 83.09 | 95.62 | 89.11 | 77.05 | 77.72 | 77.18 | 77.53 |
Precision | 95.89 | 93.30 | 84.66 | 73.57 | 76.36 | 74.31 | 82.91 | |
Recall | 72.66 | 98.71 | 97.33 | 65.50 | 63.15 | 87.78 | 74.46 | |
F1-Score | 82.65 | 95.93 | 90.55 | 69.24 | 68.91 | 80.45 | 78.41 |
Model | Metric | Anger | Anxiety | Craving | Pain | Fear | Horror | Joy | Relief | Sadness | Surprise |
---|---|---|---|---|---|---|---|---|---|---|---|
VGGNet | Accuracy | 73.87 | 86.16 | 80.73 | 82.29 | 87.52 | 79.12 | 84.21 | 81.23 | 95.22 | 70.70 |
Precision | 63.52 | 82.04 | 61.14 | 76.15 | 81.46 | 67.87 | 95.11 | 92.04 | 92.28 | 79.50 | |
Recall | 80.28 | 95.39 | 28.40 | 93.73 | 97.88 | 83.85 | 75.09 | 71.63 | 99.59 | 68.24 | |
F1-Score | 70.83 | 88.20 | 38.65 | 83.99 | 88.91 | 75.00 | 83.88 | 80.56 | 95.80 | 72.60 | |
VGGNet (p) | Accuracy | 74.81 | 83.76 | 79.57 | 80.34 | 86.38 | 78.23 | 82.12 | 77.31 | 95.16 | 72.81 |
Precision | 64.38 | 77.37 | 60.14 | 73.74 | 80.73 | 69.92 | 95.22 | 93.12 | 92.37 | 75.15 | |
Recall | 82.81 | 97.11 | 29.77 | 92.17 | 96.83 | 77.25 | 72.07 | 65.11 | 99.39 | 78.35 | |
F1-Score | 72.39 | 86.12 | 39.52 | 81.90 | 88.04 | 73.37 | 82.00 | 76.63 | 95.75 | 76.61 | |
Inception V-3 | Accuracy | 75.76 | 85.29 | 81.40 | 80.96 | 86.24 | 75.70 | 82.43 | 79.46 | 94.36 | 73.03 |
Precision | 63.38 | 80.14 | 91.79 | 73.47 | 80.12 | 62.47 | 94.86 | 92.68 | 91.84 | 73.18 | |
Recall | 92.00 | 96.93 | 14.66 | 96.47 | 97.23 | 87.80 | 71.99 | 67.74 | 98.42 | 84.89 | |
F1-Score | 75.04 | 87.73 | 25.24 | 83.41 | 87.83 | 72.95 | 81.79 | 78.18 | 95.02 | 78.47 | |
ResNet-50 | Accuracy | 72.81 | 85.38 | 79.93 | 81.21 | 86.79 | 79.12 | 85.10 | 81.79 | 94.72 | 71.48 |
Precision | 63.91 | 82.12 | 55.65 | 76.23 | 82.89 | 69.13 | 90.09 | 89.59 | 92.78 | 75.17 | |
Recall | 71.93 | 93.39 | 32.81 | 90.31 | 93.53 | 79.91 | 81.94 | 75.27 | 97.97 | 76.37 | |
F1-Score | 67.41 | 87.39 | 41.13 | 82.67 | 87.87 | 74.08 | 85.77 | 81.78 | 95.30 | 75.59 | |
ResNet-101 | Accuracy | 73.09 | 85.40 | 79.84 | 82.71 | 87.46 | 78.62 | 85.29 | 80.87 | 94.72 | 72.31 |
Precision | 63.08 | 81.02 | 55.52 | 76.32 | 83.32 | 70.46 | 93.01 | 90.90 | 92.54 | 76.84 | |
Recall | 77.40 | 95.49 | 32.30 | 94.51 | 94.40 | 74.11 | 79.30 | 72.04 | 98.27 | 75.08 | |
F1-Score | 69.49 | 87.66 | 40.72 | 84.44 | 88.50 | 72.14 | 85.52 | 80.36 | 95.32 | 75.90 | |
DenseNet | Accuracy | 73.31 | 84.88 | 80.73 | 81.10 | 87.21 | 77.95 | 82.60 | 81.26 | 93.41 | 72.17 |
Precision | 63.80 | 81.41 | 67.75 | 74.75 | 83.03 | 66.11 | 90.24 | 89.92 | 92.33 | 74.64 | |
Recall | 75.51 | 93.50 | 19.33 | 93.50 | 94.29 | 84.60 | 76.76 | 73.84 | 95.93 | 79.10 | |
F1-Score | 67.20 | 87.92 | 38.81 | 84.44 | 88.22 | 75.20 | 83.16 | 80.12 | 95.37 | 77.37 | |
EfficientNet | Accuracy | 74.46 | 86.34 | 81.40 | 83.18 | 88.12 | 77.62 | 82.72 | 78.24 | 94.91 | 72.38 |
Precision | 62.07 | 82.16 | 65.22 | 76.69 | 84.58 | 71.05 | 91.86 | 91.26 | 92.43 | 79.86 | |
Recall | 79.08 | 95.43 | 31.80 | 95.71 | 94.55 | 70.48 | 75.96 | 68.16 | 98.56 | 65.82 | |
F1-Score | 69.06 | 87.03 | 30.02 | 83.07 | 88.28 | 74.09 | 82.85 | 81.05 | 94.08 | 76.72 | |
VGGNet (P+I) | Accuracy | 75.90 | 84.24 | 79.96 | 80.43 | 87.24 | 78.23 | 82.35 | 78.32 | 95.47 | 73.56 |
Precision | 64.85 | 77.44 | 66.01 | 72.59 | 81.06 | 67.11 | 95.87 | 94.75 | 92.61 | 77.24 | |
Recall | 86.71 | 98.23 | 24.85 | 95.57 | 98.34 | 86.27 | 71.93 | 65.69 | 99.70 | 76.10 | |
F1-Score | 74.07 | 86.60 | 35.68 | 82.50 | 88.87 | 75.49 | 82.16 | 77.59 | 96.02 | 76.55 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hassan, S.Z.; Ahmad, K.; Hicks, S.; Halvorsen, P.; Al-Fuqaha, A.; Conci, N.; Riegler, M. Visual Sentiment Analysis from Disaster Images in Social Media. Sensors 2022, 22, 3628. https://doi.org/10.3390/s22103628
Hassan SZ, Ahmad K, Hicks S, Halvorsen P, Al-Fuqaha A, Conci N, Riegler M. Visual Sentiment Analysis from Disaster Images in Social Media. Sensors. 2022; 22(10):3628. https://doi.org/10.3390/s22103628
Chicago/Turabian StyleHassan, Syed Zohaib, Kashif Ahmad, Steven Hicks, Pål Halvorsen, Ala Al-Fuqaha, Nicola Conci, and Michael Riegler. 2022. "Visual Sentiment Analysis from Disaster Images in Social Media" Sensors 22, no. 10: 3628. https://doi.org/10.3390/s22103628
APA StyleHassan, S. Z., Ahmad, K., Hicks, S., Halvorsen, P., Al-Fuqaha, A., Conci, N., & Riegler, M. (2022). Visual Sentiment Analysis from Disaster Images in Social Media. Sensors, 22(10), 3628. https://doi.org/10.3390/s22103628