
Reinforcement Learning (RL) serves as a fundamental learning paradigm in the field of artificial intelligence, enabling decision-making policies through interactions with environments. However, traditional RL methods encounter challenges when dealing with large-scale or continuous state spaces due to the curse of dimensionality. Although Deep Reinforcement Learning (DRL) can handle complex environments, its lack of transparency and interpretability hinders its applicability due to the black box nature. Moreover, centralized data collection and processing methods pose privacy security risks. Federated learning offers a distributed approach that ensures privacy preservation while co-training models. However, existing federated reinforcement learning approaches have not adequately addressed communication and computation overhead issues. To address these challenges, this study proposes a tensor train decomposition-based federated reinforcement learning method that enhances efficiency and provides interpretability. By leveraging tensor to model state-action values and employing tensor decomposition techniques for dimensionality reduction, this method effectively reduces model parameters and communication overhead while maintaining strong interpretability, accelerates algorithm convergence speed. Experimental results validate the advantages of our proposed algorithm in terms of efficiency and reliability.
tensor train decomposition, federated reinforcement learning (frl), Electronic computers. Computer science, cyber-physical-social intelligence (cpsi), QA75.5-76.95
tensor train decomposition, federated reinforcement learning (frl), Electronic computers. Computer science, cyber-physical-social intelligence (cpsi), QA75.5-76.95
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
