Full Text:   <2>

CLC number: 

On-line Access: 2024-12-16

Received: 2024-05-17

Revision Accepted: 2024-09-18

Crosschecked: 0000-00-00

Cited: 0

Clicked: 3

Citations:  Bibtex RefMan EndNote GB/T7714

-   Go to

Article info.
Open peer comments

Journal of Zhejiang University SCIENCE C 1998 Vol.-1 No.-1 P.

http://doi.org/10.1631/FITEE.2400406


Significance extraction based on data augmentation for reinforcement learning


Author(s):  Yuxi HAN, Dequan LI, Yang YANG

Affiliation(s):  Faculty of Artificial Intelligence, Anhui University of Science and Technology, Huainan 232000, China

Corresponding email(s):   hanyuxi0712@163.com, leedqcpp@126.com

Key Words:  Deep reinforcement learning, Visual tasks, Generalization, Data augmentation, Significance, DeepMind Control generalization benchmark


Yuxi HAN, Dequan LI, Yang YANG. Significance extraction based on data augmentation for reinforcement learning[J]. Frontiers of Information Technology & Electronic Engineering, 1998, -1(-1): .

@article{title="Significance extraction based on data augmentation for reinforcement learning",
author="Yuxi HAN, Dequan LI, Yang YANG",
journal="Frontiers of Information Technology & Electronic Engineering",
volume="-1",
number="-1",
pages="",
year="1998",
publisher="Zhejiang University Press & Springer",
doi="10.1631/FITEE.2400406"
}

%0 Journal Article
%T Significance extraction based on data augmentation for reinforcement learning
%A Yuxi HAN
%A Dequan LI
%A Yang YANG
%J Journal of Zhejiang University SCIENCE C
%V -1
%N -1
%P
%@ 2095-9184
%D 1998
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.2400406

TY - JOUR
T1 - Significance extraction based on data augmentation for reinforcement learning
A1 - Yuxi HAN
A1 - Dequan LI
A1 - Yang YANG
J0 - Journal of Zhejiang University Science C
VL - -1
IS - -1
SP -
EP -
%@ 2095-9184
Y1 - 1998
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.2400406


Abstract: 
deep reinforcement learning has shown remarkable capabilities in visual tasks, but it does not have a good generalization ability in the context of interference signals in the input images; this approach is therefore hard to be applied to trained agents in a new environment. To enable agents to distinguish between noise signals and important pixels in images, data augmentation techniques and the establishment of auxiliary networks are proven effective solutions. We introduce a novel algorithm, namely, Saliency-extracted Q-value by augmentation (SEQA), which encourages the agent to explore unknown states more comprehensively and focus its attention on important information. Specifically, SEQA masks out interfering features and extracts salient features and then updates the mask decoder network with critic losses to encourage the agent to focus on important features and make correct decisions. We evaluate our algorithm on the DeepMind Control generalization benchmark, and the experimental results show that our algorithm greatly improves training efficiency and stability. Meanwhile, our algorithm is either superior or equal to the state-of-the-art reinforcement learning methods in terms of sample efficiency and generalization.

Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - 2024 Journal of Zhejiang University-SCIENCE