<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-17T20:37:19Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/428142" metadataPrefix="marc">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/428142</identifier><datestamp>2025-07-16T22:28:11Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452949</setSpec></header><metadata><record xmlns="http://www.loc.gov/MARC21/slim" xmlns:dcterms="http://purl.org/dc/terms/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.loc.gov/MARC21/slim http://www.loc.gov/standards/marcxml/schema/MARC21slim.xsd">
   <leader>00925njm 22002777a 4500</leader>
   <datafield ind2=" " ind1=" " tag="042">
      <subfield code="a">dc</subfield>
   </datafield>
   <datafield ind2=" " ind1=" " tag="720">
      <subfield code="a">Sharma, Robin Kumar</subfield>
      <subfield code="e">author</subfield>
   </datafield>
   <datafield ind2=" " ind1=" " tag="720">
      <subfield code="a">Casas, Marc</subfield>
      <subfield code="e">author</subfield>
   </datafield>
   <datafield ind2=" " ind1=" " tag="260">
      <subfield code="c">2023-05</subfield>
   </datafield>
   <datafield ind2=" " ind1=" " tag="520">
      <subfield code="a">Recurrent neural networks (RNNs) are widely used for&#xd;
natural language processing, time-series prediction, or text&#xd;
analysis tasks [1]. RNNs models have been widely used&#xd;
in combination with convolutional neural networks (CNNs).&#xd;
RNNs contain memory units that display dynamic and temporal&#xd;
connections between past and future data. The outstanding&#xd;
text and signal analysis properties of RNNs and other recurrent&#xd;
models like Long-Short Term Memories (LSTMs) [2] and&#xd;
Gated Recurrent Units (GRUs) [3] make them the prevalent&#xd;
choice to analyze sequential and unsegmented data like text or&#xd;
speech signals.&#xd;
RNNs have two widely used variants; one is uni-directional&#xd;
RNNs [1], which only preserves the information of the past&#xd;
because the only inputs it has seen are from the past, and the&#xd;
second is bi-directional RNNs (BRNNs) [4] which preserves&#xd;
both past and future information. The internal structure of&#xd;
RNNs and its variants inference and training in terms of data or&#xd;
control dependencies across their fundamental numerical kernels&#xd;
complicate the exploitation of model parallelism, which&#xd;
is why just data-parallelism has been traditionally applied to&#xd;
accelerate RNNs [1]. Model parallelism has not been fully&#xd;
exploited to accelerate the forward and backward propagation&#xd;
of RNNs on multi-core CPUs.&#xd;
We present two model parallelism-based approaches: WPar&#xd;
(Wavefront-Parallelization), a comprehensive approach for&#xd;
uni-directional RNNs, and B-Par (Bidirectional-Parallelization)&#xd;
for bi-directional RNNs inference and training on CPUs that&#xd;
relies on applying model parallelism into RNNs models. We&#xd;
use fine-grained pipeline parallelism in terms of tasks to&#xd;
accelerate multi-layer RNNs running on multi-core CPUs.</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">Àrees temàtiques de la UPC::Informàtica::Arquitectura de computadors</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">High performance computing</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">Deep neural network (DNN)</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">wavefront parallelization</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">task parallelism</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">recurrent neural networks (RNNs)</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">bidirectional recurrent neural networks (BRNNs)</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">long-short term memory (LSTM)</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">gated recurrent units (GRU)</subfield>
   </datafield>
   <datafield tag="653" ind2=" " ind1=" ">
      <subfield code="a">Càlcul intensiu (Informàtica)</subfield>
   </datafield>
   <datafield ind2="0" ind1="0" tag="245">
      <subfield code="a">Parallelizing recurrent neural networks and variants using OmpSs</subfield>
   </datafield>
</record></metadata></record></GetRecord></OAI-PMH>