<?xml version='1.0' encoding='UTF-8'?><?xml-stylesheet href='static/style.xsl' type='text/xsl'?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-04-21T09:40:20Z</responseDate><request verb="GetRecord" identifier="oai:ebiltegia.mondragon.edu:20.500.11984/6301" metadataPrefix="mods">https://ebiltegia.mondragon.edu/oai/request</request><GetRecord><record><header><identifier>oai:ebiltegia.mondragon.edu:20.500.11984/6301</identifier><datestamp>2024-03-22T11:44:01Z</datestamp><setSpec>com_20.500.11984_1143</setSpec><setSpec>col_20.500.11984_1148</setSpec></header><metadata><mods:mods xmlns:mods="http://www.loc.gov/mods/v3" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:doc="http://www.lyncode.com/xoai" xsi:schemaLocation="http://www.loc.gov/mods/v3 http://www.loc.gov/standards/mods/v3/mods-3-1.xsd">
   <mods:name>
      <mods:namePart>Odriozola Olalde, Haritz</mods:namePart>
   </mods:name>
   <mods:name>
      <mods:namePart>Arana-Arexolaleiba, Nestor</mods:namePart>
   </mods:name>
   <mods:extension>
      <mods:dateAvailable encoding="iso8601">2024-03-21T13:35:22Z</mods:dateAvailable>
   </mods:extension>
   <mods:extension>
      <mods:dateAccessioned encoding="iso8601">2024-03-21T13:35:22Z</mods:dateAccessioned>
   </mods:extension>
   <mods:originInfo>
      <mods:dateIssued encoding="iso8601">2023</mods:dateIssued>
   </mods:originInfo>
   <mods:identifier type="issn">1613-0073</mods:identifier>
   <mods:identifier type="other">https://katalogoa.mondragon.edu/janium-bin/janium_login_opac.pl?find&amp;ficha_no=174296</mods:identifier>
   <mods:identifier type="uri">https://hdl.handle.net/20.500.11984/6301</mods:identifier>
   <mods:abstract>Shielding methods for Reinforcement Learning agents show potential for safety-critical industrial applications. However, they still lack robustness on nominal safety, a key property for safety control systems. In the case of a significant change in the environment dynamic, shielding methods cannot guarantee safety until their inherent dynamics model is updated to the new scenario. The agent could reach risky states because the model cannot predict well. These situations could lead to catastrophic outcomes, such as damage to the cyber-physical system or loss of human lives, which are not allowed on safety-critical applications. The novel method presented in this paper, Fear Field, replicates human behaviour in those scenarios, adapting safety constraints whenever a drastic environmental change is introduced. Fear Field reduces safety violations by one order of magnitude compared to an RL agent implementing only a shield.</mods:abstract>
   <mods:language>
      <mods:languageTerm>eng</mods:languageTerm>
   </mods:language>
   <mods:accessCondition type="useAndReproduction">Attribution 4.0 International</mods:accessCondition>
   <mods:accessCondition type="useAndReproduction">http://creativecommons.org/licenses/by/4.0/</mods:accessCondition>
   <mods:accessCondition type="useAndReproduction">© 2023 The Authors</mods:accessCondition>
   <mods:subject>
      <mods:topic>Reinforcement Learning</mods:topic>
   </mods:subject>
   <mods:subject>
      <mods:topic>Shielding</mods:topic>
   </mods:subject>
   <mods:subject>
      <mods:topic>Adaptive constraints</mods:topic>
   </mods:subject>
   <mods:subject>
      <mods:topic>Robustness</mods:topic>
   </mods:subject>
   <mods:subject>
      <mods:topic>Safe AI</mods:topic>
   </mods:subject>
   <mods:subject>
      <mods:topic>ODS 9 Industria, innovación e infraestructura</mods:topic>
   </mods:subject>
   <mods:titleInfo>
      <mods:title>Fear Field: Adaptive constraints for safe environment transitions in Shielded Reinforcement Learning</mods:title>
   </mods:titleInfo>
   <mods:genre>http://purl.org/coar/resource_type/c_c94f</mods:genre>
</mods:mods></metadata></record></GetRecord></OAI-PMH>