Automatic Hierarchical Reinforcement Learning for Efficient Large-Scale Service Composition

Author(s):  
Hongbing Wang ◽  
Guicheng Huang ◽  
Qi Yu
Author(s):  
Jacob Rafati ◽  
David C. Noelle

Common approaches to Reinforcement Learning (RL) are seriously challenged by large-scale applications involving huge state spaces and sparse delayed reward feedback. Hierarchical Reinforcement Learning (HRL) methods attempt to address this scalability issue by learning action selection policies at multiple levels of temporal abstraction. Abstraction can be had by identifying a relatively small set of states that are likely to be useful as subgoals, in concert with the learning of corresponding skill policies to achieve those subgoals. Many approaches to subgoal discovery in HRL depend on the analysis of a model of the environment, but the need to learn such a model introduces its own problems of scale. Once subgoals are identified, skills may be learned through intrinsic motivation, introducing an internal reward signal marking subgoal attainment. We present a novel model-free method for subgoal discovery using incremental unsupervised learning over a small memory of the most recent experiences of the agent. When combined with an intrinsic motivation learning mechanism, this method learns subgoals and skills together, based on experiences in the environment. Thus, we offer an original approach to HRL that does not require the acquisition of a model of the environment, suitable for large-scale applications. We demonstrate the efficiency of our method on a variant of the rooms environment.


Author(s):  
Hongbing Wang ◽  
Mingzhu Gu ◽  
Qi Yu ◽  
Huanhuan Fei ◽  
Jiajie Li ◽  
...  

2019 ◽  
Vol 180 ◽  
pp. 75-90 ◽  
Author(s):  
Hongbing Wang ◽  
Mingzhu Gu ◽  
Qi Yu ◽  
Yong Tao ◽  
Jiajie Li ◽  
...  

Author(s):  
Jiang-Wen Liu ◽  
Li-Qiang Hu ◽  
Zhao-Quan Cai ◽  
Li-Ning Xing ◽  
Xu Tan

2020 ◽  
Vol 48 (3) ◽  
pp. 129-136
Author(s):  
Qihang Wu ◽  
Daifeng Li ◽  
Lu Huang ◽  
Biyun Ye

Purpose Entity relation extraction is an important research direction to obtain structured information. However, most of the current methods are to determine the relations between entities in a given sentence based on a stepwise method, seldom considering entities and relations into a unified framework. The joint learning method is an optimal solution that combines relations and entities. This paper aims to optimize hierarchical reinforcement learning framework and provide an efficient model to extract entity relation. Design/methodology/approach This paper is based on the hierarchical reinforcement learning framework of joint learning and combines the model with BERT, the best language representation model, to optimize the word embedding and encoding process. Besides, this paper adjusts some punctuation marks to make the data set more standardized, and introduces positional information to improve the performance of the model. Findings Experiments show that the model proposed in this paper outperforms the baseline model with a 13% improvement, and achieve 0.742 in F1 score in NYT10 data set. This model can effectively extract entities and relations in large-scale unstructured text and can be applied to the fields of multi-domain information retrieval, intelligent understanding and intelligent interaction. Originality/value The research provides an efficient solution for researchers in a different domain to make use of artificial intelligence (AI) technologies to process their unstructured text more accurately.


Sign in / Sign up

Export Citation Format

Share Document