<?xml version='1.0' encoding='utf-8'?>
<oai_dc:dc xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
  <dc:contributor>Christopher K. Wikle</dc:contributor>
  <dc:contributor>Mevin Hooten</dc:contributor>
  <dc:creator>Toryn L. J. Schafer</dc:creator>
  <dc:date>2022</dc:date>
  <dc:description>&lt;div&gt;&lt;p id="ID0EF" class="first"&gt;Agent-based methods allow for defining simple rules that generate complex group behaviors. The governing rules of such models are typically set a priori, and parameters are tuned from observed behavior trajectories. Instead of making simplifying assumptions across all anticipated scenarios, inverse reinforcement learning provides inference on the short-term (local) rules governing long-term behavior policies by using properties of a Markov decision process. We use the computationally efficient linearly-solvable Markov decision process to learn the local rules governing collective movement for a simulation of the selfpropelled-particle (SPP) model and a data application for a captive guppy population. The estimation of the behavioral decision costs is done in a Bayesian framework with basis function smoothing. We recover the true costs in the SPP simulation and find the guppies value collective movement more than targeted movement toward shelter.&lt;/p&gt;&lt;/div&gt;</dc:description>
  <dc:format>application/pdf</dc:format>
  <dc:identifier>10.1214/21-AOAS1529</dc:identifier>
  <dc:language>en</dc:language>
  <dc:publisher>Project Euclid</dc:publisher>
  <dc:title>Bayesian inverse reinforcement learning for collective animal movement</dc:title>
  <dc:type>article</dc:type>
</oai_dc:dc>