<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-17T20:36:01Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/410264" metadataPrefix="oai_dc">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/410264</identifier><datestamp>2025-07-22T22:49:42Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452951</setSpec></header><metadata><oai_dc:dc xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
   <dc:title>Temporally-coherent video cartoonization</dc:title>
   <dc:creator>Rayo Hernandez, Gustavo Enrique</dc:creator>
   <dc:contributor>Universitat Politècnica de Catalunya. Departament d'Arquitectura de Computadors</dc:contributor>
   <dc:contributor>Tous Liesa, Rubén</dc:contributor>
   <dc:subject>Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial</dc:subject>
   <dc:subject>Video recording</dc:subject>
   <dc:subject>Artificial intelligence</dc:subject>
   <dc:subject>Video cartoonization</dc:subject>
   <dc:subject>video-to-video translation</dc:subject>
   <dc:subject>diffusion model</dc:subject>
   <dc:subject>Stable Diffusion</dc:subject>
   <dc:subject>ControlNet</dc:subject>
   <dc:subject>EbSynth</dc:subject>
   <dc:subject>Vídeo</dc:subject>
   <dc:subject>Intel·ligència artificial</dc:subject>
   <dc:description>The automatic transformation of short background videos from real scenarios into others with a visually pleasing style, like those used in cartoons, holds application in various domains. These include animated films, video games, advertisements, and many other areas that involve visual content creation. A method or tool that can perform this task, would inspire, facilitate, and streamline the work of artists and people who produce this type of content. This thesis proposes a method that integrates multiple components to translate short background videos into others that contain a particular style. We employ Stable Diffusion, a text-to-image diffusion model, along with other technologies like ControlNet to translate keyframes from the source video, ensuring content preservation. The style of the transformed keyframes is propagated to the rest of the frames using EbSynth to make the process faster and maintain the temporal coherence. We quantitatively assess content preservation and temporal coherence using CLIP-based metrics over a new dataset of videos translated into three distinct styles. The implementation of our method is publicly available at https://github.com/gustavorayo/video-to-cartoon</dc:description>
   <dc:date>2024-01-25</dc:date>
   <dc:type>Master thesis</dc:type>
   <dc:identifier>https://hdl.handle.net/2117/410264</dc:identifier>
   <dc:identifier>183260</dc:identifier>
   <dc:language>eng</dc:language>
   <dc:rights>Open Access</dc:rights>
   <dc:format>application/pdf</dc:format>
   <dc:publisher>Universitat Politècnica de Catalunya</dc:publisher>
</oai_dc:dc></metadata></record></GetRecord></OAI-PMH>