<?xml version='1.0' encoding='UTF-8'?><?xml-stylesheet href='static/style.xsl' type='text/xsl'?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-21T04:45:24Z</responseDate><request verb="GetRecord" identifier="oai:ebiltegia.mondragon.edu:20.500.11984/6066" metadataPrefix="rdf">https://ebiltegia.mondragon.edu/oai/request</request><GetRecord><record><header><identifier>oai:ebiltegia.mondragon.edu:20.500.11984/6066</identifier><datestamp>2024-03-04T12:12:51Z</datestamp><setSpec>com_20.500.11984_473</setSpec><setSpec>col_20.500.11984_478</setSpec></header><metadata><rdf:RDF xmlns:rdf="http://www.openarchives.org/OAI/2.0/rdf/" xmlns:ow="http://www.ontoweb.org/ontology/1#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:ds="http://dspace.org/ds/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/rdf/ http://www.openarchives.org/OAI/2.0/rdf.xsd">
   <ow:Publication rdf:about="oai:ebiltegia.mondragon.edu:20.500.11984/6066">
      <dc:title>Real-Time, Model-Agnostic and User-Driven Counterfactual Explanations Using Autoencoders</dc:title>
      <dc:creator>Labaien Soto, Jokin</dc:creator>
      <dc:creator>Zugasti, Ekhi</dc:creator>
      <dc:contributor>De Carlos Garcia, Xabier</dc:contributor>
      <dc:subject>explainable AI</dc:subject>
      <dc:subject>autoencoders</dc:subject>
      <dc:subject>counterfactual explanations</dc:subject>
      <dc:description>Explainable Artificial Intelligence (XAI) has gained significant attention in recent years due to concerns over the lack of interpretability of Deep Learning models, which hinders their decision-making processes. To address this issue, counterfactual explanations have been proposed to elucidate the reasoning behind a model’s decisions by providing what-if statements as explanations. However, generating counterfactuals traditionally involves solving an optimization problem for each input, making it impractical for real-time feedback. Moreover, counterfactuals must meet specific criteria, including being user-driven, causing minimal changes, and staying within the data distribution. To overcome these challenges, a novel model-agnostic approach called Real-Time Guided Counterfactual Explanations (RTGCEx) is proposed. This approach utilizes autoencoders to generate real-time counterfactual explanations that adhere to these criteria by optimizing a multiobjective loss function. The performance of RTGCEx has been evaluated on two datasets: MNIST and Gearbox, a synthetic time series dataset. The results demonstrate that RTGCEx outperforms traditional methods in terms of speed and efficacy on MNIST, while also effectively identifying and rectifying anomalies in the Gearbox dataset, highlighting its versatility across different scenarios.</dc:description>
      <dc:date>2023-03-28T18:03:20Z</dc:date>
      <dc:date>2023-03-28T18:03:20Z</dc:date>
      <dc:date>2023</dc:date>
      <dc:type>http://purl.org/coar/resource_type/c_6501</dc:type>
      <dc:identifier>2076-3417</dc:identifier>
      <dc:identifier>https://katalogoa.mondragon.edu/janium-bin/janium_login_opac.pl?find&amp;ficha_no=172008</dc:identifier>
      <dc:identifier>https://hdl.handle.net/20.500.11984/6066</dc:identifier>
      <dc:language>eng</dc:language>
      <dc:rights>Attribution 4.0 International</dc:rights>
      <dc:rights>http://creativecommons.org/licenses/by/4.0/</dc:rights>
      <dc:rights>© 2023 The Authors</dc:rights>
      <dc:publisher>MDPI</dc:publisher>
   </ow:Publication>
</rdf:RDF></metadata></record></GetRecord></OAI-PMH>