<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-05T11:12:25Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/448978" metadataPrefix="rdf">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/448978</identifier><datestamp>2026-02-07T07:30:10Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452950</setSpec></header><metadata><rdf:RDF xmlns:rdf="http://www.openarchives.org/OAI/2.0/rdf/" xmlns:ow="http://www.ontoweb.org/ontology/1#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:ds="http://dspace.org/ds/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/rdf/ http://www.openarchives.org/OAI/2.0/rdf.xsd">
   <ow:Publication rdf:about="oai:recercat.cat:2117/448978">
      <dc:title>Empowering vector architectures for ML: The CAMP architecture for matrix multiplication</dc:title>
      <dc:creator>Esmali Nojehdeh, Mohammadreza</dc:creator>
      <dc:creator>Mokhtarnia, Hossein</dc:creator>
      <dc:creator>Pavón Rivera, Julián</dc:creator>
      <dc:creator>Rodas Quiroga, Narcís</dc:creator>
      <dc:creator>Figueras Bagué, Roger</dc:creator>
      <dc:creator>Reggiani, Enrico</dc:creator>
      <dc:creator>Moretó Planas, Miquel</dc:creator>
      <dc:creator>Unsal, Osman Sabri</dc:creator>
      <dc:creator>Cristal Kestelman, Adrián</dc:creator>
      <dc:creator>Ayguadé Parra, Eduard</dc:creator>
      <dc:subject>Àrees temàtiques de la UPC::Informàtica::Arquitectura de computadors</dc:subject>
      <dc:subject>Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Aprenentatge automàtic</dc:subject>
      <dc:subject>Vector architecture</dc:subject>
      <dc:subject>SIMD-support unit</dc:subject>
      <dc:subject>Vector processing unit</dc:subject>
      <dc:subject>Matrix multiplication</dc:subject>
      <dc:subject>Hardware-software co-design</dc:subject>
      <dc:subject>Quantization</dc:subject>
      <dc:subject>LLM</dc:subject>
      <dc:subject>CNN</dc:subject>
      <dc:description>This study presents the Cartesian Accumulative Matrix Pipeline (CAMP) architecture, a novel approach designed to enhance matrix multiplication in Vector Architectures (VAs) and Single Instruction, Multiple Data (SIMD) units. CAMP improves processing efficiency of Quantized Neural Networks (QNNs). Matrix multiplication is a cornerstone of machine learning applications, and its quantized versions are increasingly popular for more efficient operations. Unfortunately, existing VAs and SIMD-support units struggle to efficiently handle these quantized formats. In this work, we propose CAMP, a simple yet effective architecture that leverages a hybrid multiplier. The CAMP architecture significantly advances the performance of vector architectures in handling quantized data, enabling more efficient execution of matrix multiplication across various platforms, specifically targeting the ARMv8 Scalable Vector Extension (SVE) and edge RISC-V with SIMD-support unit architecture. Thanks to its hierarchical design, CAMP natively supports multiple integer precisions; here we focus on 4-bit and 8-bit, while the unit remains applicable as a general-purpose integer multiplier across other bit widths. In addition to increasing throughput, CAMP’s architectural design also contributes to energy efficiency, making it an effective solution for low-power applications. Evaluations on a range of Large Language Models (LLMs) and Convolutional Neural Networks (CNNs) demonstrate that matrix multiplication operations using the proposed micro-architecture achieve up to 17 × and 23 × performance improvements compared to their respective baselines, the ARM A64FX core and a RISC-V-based edge System-on-Chip (SoC). Furthermore, synthesis and place-and-route (PnR) of the CAMP micro-architecture — using Synopsys ICC2 for the ARM TSMC 7nm (A64FX) and Cadence Innovus for the GlobalFoundries 22nm flow (RISC-V SoC)—add only 1% and 4% area overhead, respectively, compared to the baseline designs.</dc:description>
      <dc:description>This work has received funding from the DARE SGA1 Project, from the European High-Performance Computing Joint Undertaking (JU) under Grant Agreement No. 101202459 and from PCI2024-161687-3 Project funded by MICIU/AEI/10.13039/501100011033 and European’s Union “NextGenerationEU”/PRTR. The JU receives support from the European Union’s Horizon Europe research and innovation programme and Spain, Germany, Czechia, Italy, Netherlands, Belgium, Finland, Greece, Croatia, Portugal, Poland, Sweden, France and Austria. This research was supported by the Spanish Ministry of Science and Innovation through contract PID2023-146511NB-I00, PID2023-147979NB-C21,and by the Ministry for Digital Transformation and Public Service, within the framework of the Recovery, Transformation and Resilience Plan – NextGenerationEU REGAGE22e00058408992; the Generalitat of Catalunya through contract 2021-SGR-00763; and the Lenovo-BSC Framework Contract (2020).</dc:description>
      <dc:description>Peer Reviewed</dc:description>
      <dc:description>Postprint (published version)</dc:description>
      <dc:date>2025</dc:date>
      <dc:type>Conference report</dc:type>
      <dc:relation>https://dl.acm.org/doi/10.1145/3725843.3760547</dc:relation>
      <dc:relation>info:eu-repo/grantAgreement/AEI//PCI2024-161687-3</dc:relation>
      <dc:rights>http://creativecommons.org/licenses/by/4.0/</dc:rights>
      <dc:rights>Open Access</dc:rights>
      <dc:rights>Attribution 4.0 International</dc:rights>
      <dc:publisher>Association for Computing Machinery (ACM)</dc:publisher>
   </ow:Publication>
</rdf:RDF></metadata></record></GetRecord></OAI-PMH>