Del píxel a las resonancias visuales: la imagen con voz propia

Publication date

2016-07-25T13:30:42Z

2016-07-25T13:30:42Z

2016-06

2016-07-25T13:30:47Z

Abstract

The objective of our research is to develop a series of computer vision programs to search for analogies in large datasets¿in this case, collections of images of abstract paintings¿ based solely on their visual content without textual annotation. We have programmed an algorithm based on a specific model of image description used in computer vision. This approach involves placing a regular grid over the image and selecting a pixel region around each node. Dense features computed over this regular grid with overlapping patches are used to represent the images. Analysing the distances between the whole set of image descriptors we are able to group them according to their similarity and each resulting group will determines what we call 'visual words'. This model is called Bag-of-Words representation Given the frequency with which each visual word occurs in each image, we apply the method pLSA (Probabilistic Latent Semantic Analysis), a statistical model that classifies fully automatically, without any textual annotation, images according to their formal patterns. In

Document Type

Article


Published version

Language

Spanish

Publisher

Euskal Herriko Unibertsitateko Argitalpen Zerbitzua

Related items

Reproducció del document publicat a: http://www.ehu.eus/ojs/index.php/ausart/article/view/16670/14642

AusArt. Journal for Research in Art, 2016, vol. 4, num. 1, p. 19-28

Recommended citation

This citation was generated automatically.

Rights

cc-by-sa (c) Rosado Rodrigo, Pilar et al., 2016

http://creativecommons.org/licenses/by-sa/3.0/es

This item appears in the following Collection(s)