Service composition based on multi-agent in the cooperative game

2017 
Abstract The principle of service composition based on multi-agent is that multi-agent can coordinate to reach Pareto-optimal Nash equilibrium. Reinforcement learning algorithms can be used to deal with the coordination problem in cooperative games. In this paper, the multi-agent coordination problems in cooperative games for different user preference is investigated. In our case, each agent can represent a user’s preference, and it finally learns a policy that is best fit for that user. Most previous works study the deterministic gain of a state. However, in practical service environments, the gain may be nondeterministic due to unstable Quality of Service (QoS). In addition, user preference should be considered. To avoid local optimal solution, we let each agent randomly change interacting partners in each iteration. Thus, an agent can learn its optimal strategy by interacting repeatedly with the rest of agents representing different user preference. The experimental results show that our reinforcement learning algorithm outperforms other learning methods.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    18
    References
    6
    Citations
    NaN
    KQI
    []