This paper proposes an adaptive method to enable imitation learning from expert demonstrations in a multi-agent context. Our work employs the inverse reinforcement learning method to a coupled Dynamic Bayesian Network to facilitate dynamic learning in an interactive system. This method studies the interaction at both discrete and continuous levels by identifying inter-relationships between the objects to facilitates the prediction of an expert agent’s demonstrations. We evaluate the learning procedure in the scene of learner agent based on probabilistic reward function. Our goal is to estimate policies that predicted trajectories match the observed one by minimizing the Kullback- Leiber divergence. The reward policies provide a probabilistic dynamic structure to minimize the abnormalities.
Observational Learning: Imitation Through an Adaptive Probabilistic Approach
Sheida Nozari;Lucio Marcenaro;Carlo Regazzoni
2021-01-01
Abstract
This paper proposes an adaptive method to enable imitation learning from expert demonstrations in a multi-agent context. Our work employs the inverse reinforcement learning method to a coupled Dynamic Bayesian Network to facilitate dynamic learning in an interactive system. This method studies the interaction at both discrete and continuous levels by identifying inter-relationships between the objects to facilitates the prediction of an expert agent’s demonstrations. We evaluate the learning procedure in the scene of learner agent based on probabilistic reward function. Our goal is to estimate policies that predicted trajectories match the observed one by minimizing the Kullback- Leiber divergence. The reward policies provide a probabilistic dynamic structure to minimize the abnormalities.File | Dimensione | Formato | |
---|---|---|---|
ICAS2021 Paper Imitation learning.pdf
accesso aperto
Descrizione: Contributo in atti di convegno
Tipologia:
Documento in Post-print
Dimensione
686.79 kB
Formato
Adobe PDF
|
686.79 kB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.