Methodologies and Applications of Image Understanding in Cultural and Artistic Heritage

A special issue of Journal of Imaging (ISSN 2313-433X). This special issue belongs to the section "Computer Vision and Pattern Recognition".

Deadline for manuscript submissions: 31 October 2024 | Viewed by 2933

Special Issue Editors


E-Mail Website
Guest Editor
Digital Society Initiative, University of Zurich, Rämistrasse 69, 8001 Zürich, Switzerland
Interests: digital arts and humanities; computer vision; multimodal deep learning; explainable and human-centered AI

E-Mail Website
Guest Editor
Department of Environmental Sciences, Informatics and Statistics, Ca’ Foscari University of Venice, Via Torino, 155, Mestre, 30172 Venezia, VE, Italy
Interests: computer vision; deep learning; image analysis of visual art and cultural heritage

Special Issue Information

Dear Colleagues,

This Special Issue is dedicated to the various challenges and possibilities, both in terms of methodology and application, that emerge when computational methods are used to study, analyze, and interpret images in the context of art and culture. 

What does it mean to understand an image? The application of computational methods, such as object detection, segmentation, or classification, to images is usually motivated by the need to solve a particular problem. In the context of computer vision and deep learning, these methods are usually applied to general-purpose image datasets (ImageNet, MS Coco, etc.) or domain-specific image datasets (medical, satellite, traffic, etc.). Images related to art or cultural heritage may also represent another type of domain-specific image dataset. However, understanding an image becomes a much more complex task if we aim to integrate computational approaches of image understanding with the methodological and theoretical frameworks of traditional disciplines dedicated to the study of images in art and culture (e.g., art history and cultural and visual studies).

This Special Issue invites contributions that address different aspects of image understanding with the common goal of bridging cross-disciplinary gaps.

The topics of interest include, but are not limited to, the following:

  • Computational analysis of images as visual cultural artifacts;
  • Contextually meaningful image searches and similarity retrieval in artwork collections;
  • Deep learning-based approaches for studying artwork images;
  • Computer vision applications in cultural heritage;
  • Multimodal deep learning for image understanding;
  • Understanding AI-generated images in the context of art and culture.

Dr. Eva Cetinic
Dr. Sinem Aslan
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at mdpi.longhoe.net by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Journal of Imaging is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • computer vision
  • deep learning
  • artwork analysis
  • cultural heritage
  • generative AI
  • multimodality
  • cultural AI
  • art datasets
  • digital art history
  • digital visual studies

Published Papers (3 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

21 pages, 11698 KiB  
Article
GOYA: Leveraging Generative Art for Content-Style Disentanglement
by Yankun Wu, Yuta Nakashima and Noa Garcia
J. Imaging 2024, 10(7), 156; https://doi.org/10.3390/jimaging10070156 - 26 Jun 2024
Viewed by 380
Abstract
The content-style duality is a fundamental element in art. These two dimensions can be easily differentiated by humans: content refers to the objects and concepts in an artwork, and style to the way it looks. Yet, we have not found a way to [...] Read more.
The content-style duality is a fundamental element in art. These two dimensions can be easily differentiated by humans: content refers to the objects and concepts in an artwork, and style to the way it looks. Yet, we have not found a way to fully capture this duality with visual representations. While style transfer captures the visual appearance of a single artwork, it fails to generalize to larger sets. Similarly, supervised classification-based methods are impractical since the perception of style lies on a spectrum and not on categorical labels. We thus present GOYA, which captures the artistic knowledge of a cutting-edge generative model for disentangling content and style in art. Experiments show that GOYA explicitly learns to represent the two artistic dimensions (content and style) of the original artistic image, paving the way for leveraging generative models in art analysis. Full article
12 pages, 7865 KiB  
Article
Weakly Supervised SVM-Enhanced SAM Pipeline for Stone-by-Stone Segmentation of the Masonry of the Loire Valley Castles
by Stuardo Lucho, Sylvie Treuillet, Xavier Desquesnes, Remy Leconge and Xavier Brunetaud
J. Imaging 2024, 10(6), 148; https://doi.org/10.3390/jimaging10060148 - 19 Jun 2024
Viewed by 322
Abstract
The preservation of historical monuments presents a formidable challenge, particularly in monitoring the deterioration of building materials over time. Chateau de Chambord’s facade suffers from common issues such as flaking and spalling, which require meticulous stone and joint map** from experts manually for [...] Read more.
The preservation of historical monuments presents a formidable challenge, particularly in monitoring the deterioration of building materials over time. Chateau de Chambord’s facade suffers from common issues such as flaking and spalling, which require meticulous stone and joint map** from experts manually for restoration efforts. Advancements in computer vision have allowed machine-learning models to help in the automatic segmentation process. In this research, a custom architecture defined as SAM-SVM is proposed, to perform stone segmentation, based on the Segment Anything Model (SAM) and Support Vector Machines (SVM). By exploiting the zero-shot learning capabilities of SAM and its customizable input parameters, we obtain segmentation mask for stones and joints, which are then classified using SVM. Two more SAMs (three in total) are used, depending on how many stones are left to segment. Through extensive experimentation and evaluation, supported by computer vision methods, the proposed architecture achieves a Dice coefficient of 85%. Our results highlight the potential of SAM in cultural heritage conservation, providing a scalable and efficient solution for stone segmentation in historic monuments. This research contributes valuable insights and methodologies to the ongoing conservation efforts of Château de Chambord and could be extrapolated to other monuments. Full article
Show Figures

Figure 1

20 pages, 25770 KiB  
Article
Exploring Emotional Stimuli Detection in Artworks: A Benchmark Dataset and Baselines Evaluation
by Tianwei Chen, Noa Garcia, Liangzhi Li and Yuta Nakashima
J. Imaging 2024, 10(6), 136; https://doi.org/10.3390/jimaging10060136 - 4 Jun 2024
Viewed by 601
Abstract
We introduce an emotional stimuli detection task that targets extracting emotional regions that evoke people’s emotions (i.e., emotional stimuli) in artworks. This task offers new challenges to the community because of the diversity of artwork styles and the subjectivity of emotions, which can [...] Read more.
We introduce an emotional stimuli detection task that targets extracting emotional regions that evoke people’s emotions (i.e., emotional stimuli) in artworks. This task offers new challenges to the community because of the diversity of artwork styles and the subjectivity of emotions, which can be a suitable testbed for benchmarking the capability of the current neural networks to deal with human emotion. For this task, we construct a dataset called APOLO for quantifying emotional stimuli detection performance in artworks by crowd-sourcing pixel-level annotation of emotional stimuli. APOLO contains 6781 emotional stimuli in 4718 artworks for validation and testing. We also evaluate eight baseline methods, including a dedicated one, to show the difficulties of the task and the limitations of the current techniques through qualitative and quantitative experiments. Full article
Show Figures

Figure 1

Back to TopTop