<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-13T13:29:35Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/376954" metadataPrefix="didl">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/376954</identifier><datestamp>2025-07-23T01:09:56Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452951</setSpec></header><metadata><d:DIDL xmlns:d="urn:mpeg:mpeg21:2002:02-DIDL-NS" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="urn:mpeg:mpeg21:2002:02-DIDL-NS http://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-21_schema_files/did/didl.xsd">
   <d:Item id="hdl_2117_376954">
      <d:Descriptor>
         <d:Statement mimeType="application/xml; charset=utf-8">
            <dii:Identifier xmlns:dii="urn:mpeg:mpeg21:2002:01-DII-NS" xsi:schemaLocation="urn:mpeg:mpeg21:2002:01-DII-NS http://standards.iso.org/ittf/PubliclyAvailableStandards/MPEG-21_schema_files/dii/dii.xsd">urn:hdl:2117/376954</dii:Identifier>
         </d:Statement>
      </d:Descriptor>
      <d:Descriptor>
         <d:Statement mimeType="application/xml; charset=utf-8">
            <oai_dc:dc xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:dc="http://purl.org/dc/elements/1.1/" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
               <dc:title>Understanding and improving self-attention mechanisms</dc:title>
               <dc:creator>Pujol Perich, David</dc:creator>
               <dc:subject>Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Llenguatge natural</dc:subject>
               <dc:subject>Deep learning (Machine learning)</dc:subject>
               <dc:subject>Natural language processing (Computer science)</dc:subject>
               <dc:subject>Computer vision</dc:subject>
               <dc:subject>aprenentatge profund</dc:subject>
               <dc:subject>self-attention</dc:subject>
               <dc:subject>processament de llenguatge natural</dc:subject>
               <dc:subject>visió per computació</dc:subject>
               <dc:subject>models d'ordre superior</dc:subject>
               <dc:subject>deep learning</dc:subject>
               <dc:subject>efficient Transformers</dc:subject>
               <dc:subject>natural language processing</dc:subject>
               <dc:subject>computer vision</dc:subject>
               <dc:subject>high-order models</dc:subject>
               <dc:subject>aprenentatge automàtic</dc:subject>
               <dc:subject>machine learning</dc:subject>
               <dc:subject>Transformers</dc:subject>
               <dc:subject>Aprenentatge profund</dc:subject>
               <dc:subject>Tractament del llenguatge natural (Informàtica)</dc:subject>
               <dc:subject>Visió per ordinador</dc:subject>
               <dc:description>Recent years have seen the vast potential of the Transformer model, as it is arguably the first general-purpose architecture in the sense that achieves state-of-the-art performance in numerous fields –e.g., Computer Vision, Natural Language Processing, Autonomous driving– using minimal architecture modifications. The success of Transformers greatly relies on the use of the self-attention mechanism, the understanding of which still remains somewhat obscure. In this thesis, we first focus on bridging this gap from an empirical perspective, studying its main inductive biases and limitations. We also propose the ReLA Nyströmformer, a novel architecture that attains a linear complexity –considerably improving the original quadratic complexity of the self-attention– meanwhile proving to be empirically superior to a vast set of state-of-the-art benchmarks. We also notice that Transformer architectures are ultimately third-order-interaction-based models –i.e., can be formalized as third-order polynomials– which tractability strongly depends on a number of inductive biases. This motivates the last part of this thesis where we discuss the suitability of devising higher-order models –i.e., based on higher-order polynomials– both from a predictive and interpretability point of view. Finally, we propose two novel architectures, the Low-rank Deep Polynomial Network and the Adaptive attention, based on low-rank projections and automatic attention pattern learning. The state-of- the-art performance of both of these models underpins the need of researching further in this direction to fully elucidate their potential.</dc:description>
               <dc:date>2022-07-10</dc:date>
               <dc:type>Master thesis</dc:type>
               <dc:rights>Restricted access - author's decision</dc:rights>
               <dc:publisher>Universitat Politècnica de Catalunya</dc:publisher>
            </oai_dc:dc>
         </d:Statement>
      </d:Descriptor>
   </d:Item>
</d:DIDL></metadata></record></GetRecord></OAI-PMH>