Del píxel a las resonancias visuales: la imagen con voz propia

Fecha de publicación

2016-07-25T13:30:42Z

2016-07-25T13:30:42Z

2016-06

2016-07-25T13:30:47Z

Resumen

The objective of our research is to develop a series of computer vision programs to search for analogies in large datasets¿in this case, collections of images of abstract paintings¿ based solely on their visual content without textual annotation. We have programmed an algorithm based on a specific model of image description used in computer vision. This approach involves placing a regular grid over the image and selecting a pixel region around each node. Dense features computed over this regular grid with overlapping patches are used to represent the images. Analysing the distances between the whole set of image descriptors we are able to group them according to their similarity and each resulting group will determines what we call 'visual words'. This model is called Bag-of-Words representation Given the frequency with which each visual word occurs in each image, we apply the method pLSA (Probabilistic Latent Semantic Analysis), a statistical model that classifies fully automatically, without any textual annotation, images according to their formal patterns. In

Tipo de documento

Artículo


Versión publicada

Lengua

Castellano

Publicado por

Euskal Herriko Unibertsitateko Argitalpen Zerbitzua

Documentos relacionados

Reproducció del document publicat a: http://www.ehu.eus/ojs/index.php/ausart/article/view/16670/14642

AusArt. Journal for Research in Art, 2016, vol. 4, num. 1, p. 19-28

Citación recomendada

Esta citación se ha generado automáticamente.

Derechos

cc-by-sa (c) Rosado Rodrigo, Pilar et al., 2016

http://creativecommons.org/licenses/by-sa/3.0/es

Este ítem aparece en la(s) siguiente(s) colección(ones)