Skip to main navigation Skip to search Skip to main content

QoE-Driven Content-Centric Caching with Deep Reinforcement Learning in Edge-Enabled IoT

Research output: Contribution to journalArticlepeer-review

61 Scopus citations

Abstract

When humans learn several skills to solve multiple tasks, they exhibit an extraordinary capacity to transfer knowledge between them. The authors present here the last enhanced version of a bioinspired reinforcement-learning (RL) modular architecture able to perform skill-to-skill knowledge transfer and called transfer expert RL (TERL) model. TERL architecture is based on a RL actor-critic model where both actor and critic have a hierarchical structure, inspired by the mixture-of-experts model, formed by a gating network that selects experts specializing in learning the policies or value functions of different tasks. A key feature of TERL is the capacity of its gating networks to accumulate, in parallel, evidence on the capacity of experts to solve the new tasks so as to increase the responsibility for action of the best ones. A second key feature is the use of two different responsibility signals for the experts' functioning and learning: This allows the training of multiple experts for each task so that some of them can be later recruited to solve new tasks and avoid catastrophic interference. The utility of TERL mechanisms is shown with tests involving two simulated dynamic robot arms engaged in solving reaching tasks, in particular a planar 2-DoF arm, and a 3-D 4-DoF arm.

Original languageEnglish
Article number8871365
Pages (from-to)12-20
Number of pages9
JournalIEEE Computational Intelligence Magazine
Volume14
Issue number4
DOIs
StatePublished - Nov 2019

Fingerprint

Dive into the research topics of 'QoE-Driven Content-Centric Caching with Deep Reinforcement Learning in Edge-Enabled IoT'. Together they form a unique fingerprint.

Cite this