<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-18T05:17:36Z</responseDate><request verb="GetRecord" identifier="oai:www.recercat.cat:2117/180457" metadataPrefix="oai_dc">https://recercat.cat/oai/request</request><GetRecord><record><header><identifier>oai:recercat.cat:2117/180457</identifier><datestamp>2026-02-07T09:30:50Z</datestamp><setSpec>com_2072_1033</setSpec><setSpec>col_2072_452950</setSpec></header><metadata><oai_dc:dc xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
   <dc:title>Time-domain speech enhancement using generative adversarial networks</dc:title>
   <dc:creator>Pascual de la Puente, Santiago</dc:creator>
   <dc:creator>Serra, Joan</dc:creator>
   <dc:creator>Bonafonte Cávez, Antonio</dc:creator>
   <dc:contributor>Universitat Politècnica de Catalunya. Doctorat en Teoria del Senyal i Comunicacions</dc:contributor>
   <dc:contributor>Universitat Politècnica de Catalunya. Departament de Teoria del Senyal i Comunicacions</dc:contributor>
   <dc:contributor>Universitat Politècnica de Catalunya. VEU - Grup de Tractament de la Parla</dc:contributor>
   <dc:subject>Àrees temàtiques de la UPC::Enginyeria de la telecomunicació</dc:subject>
   <dc:subject>Speech processing systems</dc:subject>
   <dc:subject>Neural networks (Computer science)</dc:subject>
   <dc:subject>Speech enhancement</dc:subject>
   <dc:subject>Audio transformation</dc:subject>
   <dc:subject>Generative adversarial network</dc:subject>
   <dc:subject>Neural networks</dc:subject>
   <dc:subject>Processament de la parla</dc:subject>
   <dc:subject>Reconeixement automàtic de la parla</dc:subject>
   <dc:subject>Xarxes neuronals (Informàtica)</dc:subject>
   <dc:description>Speech enhancement improves recorded voice utterances to eliminate noise that might be impeding their intelligibility or compromising their quality. Typical speech enhancement systems are based on regression approaches that subtract noise or predict clean signals. Most of them do not operate directly on waveforms. In this work, we propose a generative approach to regenerate corrupted signals into a clean version by using generative adversarial networks on the raw signal. We also explore several variations of the proposed system, obtaining insights into proper architectural choices for an adversarially trained, convolutional autoencoder applied to speech. We conduct both objective and subjective evaluations to assess the performance of the proposed method. The former helps us choose among variations and better tune hyperparameters, while the latter is used in a listening experiment with 42 subjects, confirming the effectiveness of the approach in the real world. We also demonstrate the applicability of the approach for more generalized speech enhancement, where we have to regenerate voices from whispered signals.</dc:description>
   <dc:description>Peer Reviewed</dc:description>
   <dc:description>Postprint (author's final draft)</dc:description>
   <dc:date>2019-11-01</dc:date>
   <dc:type>Article</dc:type>
   <dc:identifier>Pascual, S.; Serra, J.; Bonafonte, A. Time-domain speech enhancement using generative adversarial networks. "Speech communication", 1 Novembre 2019, vol. 114, p. 10-21.</dc:identifier>
   <dc:identifier>0167-6393</dc:identifier>
   <dc:identifier>https://hdl.handle.net/2117/180457</dc:identifier>
   <dc:identifier>10.1016/j.specom.2019.09.001</dc:identifier>
   <dc:language>eng</dc:language>
   <dc:relation>https://www.sciencedirect.com/science/article/abs/pii/S0167639319301359</dc:relation>
   <dc:relation>info:eu-repo/grantAgreement/MINECO//TEC2015-69266-P/ES/TECNOLOGIAS DE APRENDIZAJE PROFUNDO APLICADAS AL PROCESADO DE VOZ Y AUDIO/</dc:relation>
   <dc:rights>Open Access</dc:rights>
   <dc:format>12 p.</dc:format>
   <dc:format>application/pdf</dc:format>
</oai_dc:dc></metadata></record></GetRecord></OAI-PMH>