Recently I described simple K-bandit problem and solution. I also did a little introduction to Reinforcement Learning problem. Today I am still going to focus on the same problem with a little bit more terminology and few different algorithms (or more like few different variants). I am not going to exhaust the topic as it’s pretty broad and well studied but to give myself and you dear reader some overview on it. Let’s begin. 🏁
Stationary vs Non-stationary
In the last post, I presented one simple algorithm in 2 variants. Greedy and Ɛ-greedy ones to solve our K-armed bandit problem. It was an easy mode which was good as an introduction to the subject and not so good as a real-life-like problem. The data which our algorithm operated on was stationary. In plain words:
Stationary problem – there is always the best answer and it never changes
It doesn’t work that way in a real world! 🌏 Like a data of users, behavior might be different at night or when some event happens. Or the next game level might have some additional obstacle. We would like our algorithms to fins such subtleties.
Non-stationary – there is always the best answer but it could change any time
I had to adjust a little bit my problem and reward generation functions to accommodate the non-stationary problem. And as I plot a lot this time I needed to make plotting and comparing stuff easier.
What I did was wrapping everything in class to easily pass problem object to a solution object and use it. Really only new thing is a
stationary flag. If it’s
false then I add to the
values normally distributed random numbers with small variance.
optimal property is there to count performance of the solution. Yeah, it might be an overkill but I am a software engineer by trade and I just like writing code and optimizing it. I also created a base class for solutions to count the data I needed for plotting charts.
I wrote two helper functions. One to generate average data over X runs and the other to create charts. My jupyter-notebook is a mess right now but I will publish it when I finish having fun with K-armed bandit.
Ɛ-greedy solution to a new problem
Long story short. I had to modify a previous solution to accommodate class-based approach. The really new thing is
initial_value needed later, for now, it will be always set to
And here is its performance on our new problem:
In this kind of problem the more our solution tries to explore the better it performs. On a stationary problem chart, the lines were much closer together.
As the data in our environment change, it’s short-sighted to assume every data-point has exactly the same value. And that’s what we did up until now. Only point #9 had changed.
In Ɛ-greedy we calculated mean to estimate value function (
Q) here we use a weighted average. I parametrized only the
exploration_rate is at
0.1. In Reinforcement Learning: An Introduction all maths are nicely described. The update rule in point #9 is now calculating the weighted average having only previous weighted average value and new value so we don’t have to store all values. Check previous post for why. If you don’t believe the math here go to comments or to the book.
My goal wasn’t to find the optimal solution but to reproduce algorithms with some personal flair. We can see the results are pretty much on par with the first one. So our new code doesn’t seem to perform much better the last. It’s not useless either, results are quite on par and the chart behaves differently.
Optimistic vs Realistic
Up until this point, all our algorithms were realistic. That’s what the
initial_value parameter is for. The realistic algorithm assumes that the all possible choices at the beginning are not either good (positive value) or bad (negative value). They are just zero and it will figure out on the fly the estimations.
The optimistic solution assumes that all actions could be splendid. That means significantly higher than the real possible values. That gives algorithm an incentive to try it all. Let’s see how it compares.
My conclusion is following. The realistic algorithm starts stronger as it sticks to action more when it finds a good one. But the optimistic catches up. In a long run, optimistic algorithms are (almost) always winning. Life lesson: it’s good to be optimist.
Upper-Confidence-Bound (UCB) action selection
But can we do better? Sure! By changing not value function estimation calculation but only the action selection method. What this method does is calculate which action to take not by how recent action was. This is the way we did it before. UCB judges an action by its potential. Let’s calculate it first.
Sorry if my math to plain words formula is unclear. We figure out values for each action by adding some things to our actual estimate of action value –
Q. Those things are:
c– exploration parameter, it’s just a hyperparameter, which means we have to optimize it ourselves.
cshould be bigger than 0
all actions taken– a count of all actions. Or in other words a current step number
action count– how many times we took this specific action.
When you take some action your
action count grows.
all actions taken grows too, but since it’s in a natural logarithm the numerator doesn’t increase as fast. When denominator grows the whole term (which is called uncertainty term) is smaller. Plain 🇬🇧: The more we explore given action, the more certain we are about it and the less certain we are about other actions. It changes for all actions so it makes sure that all actions will be explored. Given enough of steps of course. Here is an actual code:
In point #4 we count UCB dictionary which will our decision be based on.
count_ucb function calculates UCB for one action. Dictionary comprehension in #4 makes estimates for whole
Q dictionary. Then we take an option with maximum value. Note: It’s important to use big value (like
sys.maxsize in my case) when the action had never been taken for everything to work correctly. Again more math in the Reinforcement Learning: An Introduction. Let’s see the results.
Nice! If you remember the values for other solutions you can see it’s way bigger. So it looks like we nailed it. But keep in mind that UCB works only well for K-armed bandit problems. It doesn’t generalize to other reinforcement learning problems. 😞
Now it’s the time to look at how every solution compares. Here is my totally non-scientific competition with all algorithms described today.
UCB is a clear winner with both optimistic approaches on a 2nd and 3rd. The realistic one trails on 1000 of steps.
What I hoped to achieve with is post is improve my understanding of Reinforcement Learning further by learning basics. I hope you will find it useful. I had my share of fun. 😄