Del píxel a las resonancias visuales: la imagen con voz propia

Data de publicació

2016-07-25T13:30:42Z

2016-07-25T13:30:42Z

2016-06

2016-07-25T13:30:47Z

Resum

The objective of our research is to develop a series of computer vision programs to search for analogies in large datasets¿in this case, collections of images of abstract paintings¿ based solely on their visual content without textual annotation. We have programmed an algorithm based on a specific model of image description used in computer vision. This approach involves placing a regular grid over the image and selecting a pixel region around each node. Dense features computed over this regular grid with overlapping patches are used to represent the images. Analysing the distances between the whole set of image descriptors we are able to group them according to their similarity and each resulting group will determines what we call 'visual words'. This model is called Bag-of-Words representation Given the frequency with which each visual word occurs in each image, we apply the method pLSA (Probabilistic Latent Semantic Analysis), a statistical model that classifies fully automatically, without any textual annotation, images according to their formal patterns. In

Tipus de document

Article


Versió publicada

Llengua

Castellà

Publicat per

Euskal Herriko Unibertsitateko Argitalpen Zerbitzua

Documents relacionats

Reproducció del document publicat a: http://www.ehu.eus/ojs/index.php/ausart/article/view/16670/14642

AusArt. Journal for Research in Art, 2016, vol. 4, num. 1, p. 19-28

Citació recomanada

Aquesta citació s'ha generat automàticament.

Drets

cc-by-sa (c) Rosado Rodrigo, Pilar et al., 2016

http://creativecommons.org/licenses/by-sa/3.0/es

Aquest element apareix en la col·lecció o col·leccions següent(s)