Customized Action Selection in RL DQN

2 views (last 30 days)
ches
ches on 11 Jan 2021
Edited: ches on 20 Jan 2021
Hi,
I would like to ask if the latest Reinforcement Learning (RL) toolbox version supports customized action selection.
I’m currently using a DQN agent, and the action in each time step is selected randomly following the epsilon-greedy algorithm. However, I would like to feed in some probabilities in the action selection, such that certain actions are more likely to be chosen. Is this possible using the RL toolbox?
Thank you!

Answers (1)

Emmanouil Tzorakoleftherakis
Edited: Emmanouil Tzorakoleftherakis on 16 Jan 2021
Hello,
I believe this is not possible yet. A potential workaround (although not state dependent) would be to emulate a pdf by providing actions with higher probabilities multiple times when creating your action space with rlFinitesetSpec but I haven't tested that. So something like:
actInfo = rlFiniteSetSpec([-2 0 2 2 2])
  1 Comment
ches
ches on 20 Jan 2021
Edited: ches on 20 Jan 2021
Hello,
Thank you for the information.
I'm currently trying to improve the exploration during training, so I'm thinking of other ways to do that apart from adjusting the epsilon parameters of the epsilon-greedy algorithm.
In line with that, may I also ask if the following are possible in the latest RL toolbox?
- Setting optimistic initial values
- Other exploration strategies (such as Boltzmann)
Thanks!

Sign in to comment.

Products


Release

R2020b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!