<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-17T11:29:01Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/117560" metadataPrefix="oai_dc">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/117560</identifier><datestamp>2025-07-22T17:18:16Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452951</setSpec></header><metadata><oai_dc:dc xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
   <dc:title>Online action detection</dc:title>
   <dc:creator>Pan, Junting</dc:creator>
   <dc:contributor>Giró Nieto, Xavier</dc:contributor>
   <dc:contributor>Chang, Shih-Fu</dc:contributor>
   <dc:subject>Àrees temàtiques de la UPC::Enginyeria de la telecomunicació</dc:subject>
   <dc:subject>Machine learning</dc:subject>
   <dc:subject>Neural networks (Computer science)</dc:subject>
   <dc:subject>Video analysis</dc:subject>
   <dc:subject>Online action detection</dc:subject>
   <dc:subject>Convolutional neural networks</dc:subject>
   <dc:subject>Deep learning</dc:subject>
   <dc:subject>Generative adversarial networks</dc:subject>
   <dc:subject>Aprenentatge automàtic</dc:subject>
   <dc:subject>Xarxes neuronals (Informàtica)</dc:subject>
   <dc:description>The details of the work will be defined when the student reaches his destination.</dc:description>
   <dc:description>In online detection, the objective is to detect the start of an action in a video stream as soon as it happens. It is an important yet challenging problem. In many realistic scenarios, we need to detect the action before the action is completed. For example, in the autonomous driving system, it is crucial to detect whether the pedestrian is crossing the street well in time in order to make a decision to stop or to reduce the velocity. Online action detection is a very challenging task in many aspects. It is very hard to predict the start of action for three reasons: First, the background is very diverse. Moreover, there is only a few action instance in a very long video. Last but not least, the model only observes part of the action to predict. To address those challenges, we propose a framework for online action detection and simulate experiments on a large-scale untrimmed video dataset. With the proposed method we have obtained very competitive performance. We also proposed a new evaluation metric for online detection models: Point mean Average Precision (Point mAP), a more appropriate metric than the existing evaluation metrics that have been designed for action detection in an offline setting. We have conducted experiments on THUMOS'14 dataset of video analysis where our proposed model achieved the state-of-the-art performance on the online action detection task.</dc:description>
   <dc:date>2017-10</dc:date>
   <dc:type>Master thesis</dc:type>
   <dc:identifier>https://hdl.handle.net/2117/117560</dc:identifier>
   <dc:identifier>ETSETB-230.127224</dc:identifier>
   <dc:language>eng</dc:language>
   <dc:rights>S'autoritza la difusió de l'obra mitjançant la llicència Creative Commons o similar 'Reconeixement-NoComercial- SenseObraDerivada'</dc:rights>
   <dc:rights>http://creativecommons.org/licenses/by-nc-nd/3.0/es/</dc:rights>
   <dc:rights>Open Access</dc:rights>
   <dc:format>application/pdf</dc:format>
   <dc:publisher>Universitat Politècnica de Catalunya</dc:publisher>
</oai_dc:dc></metadata></record></GetRecord></OAI-PMH>