Skip to content

The-Faust/bandit_agents

Repository files navigation

Bandit Agents Library (A concise and easy to use library to help with k-armed bandit problems)

This library is an attempt at creating a library that would streamline the resolution of k-armed bandit problems

It was mostly inspired by the first few chapters of Reinforcement Learning: An Introduction

installation

There are a few ways to install the library

  • pip: pip install https://github.com/The-Faust/bandit_agents
  • local: clone or download a release of the library and add it to your python path

Note: Once I have a decent documentation going on I intend on putting the library in pypi and conda-forge

development installation

If you intend on forking the project or you are a contributor working with me you will have to setup the development environment

requirements for installation of the dev env

To make use of the setup script you must have anaconda3 installed and have conda-lock installed in the environment in which you will execute setup_dev_env.sh

setup script

to execute the setup script execute the following

chmod +x setup_dev_env.sh
./setup_dev_env.sh

executing setup_dev_env.sh will do the following:

  1. give the roght permissions to build_env.sh and compile_env_lock.sh
  2. if lock file does not exists; execute compile_env_lock.sh
  3. create the conda environement by executing build_env.sh
  4. install the pre-commit hooks found in .pre-commit-config.yaml

You can then activate the environement in conda: conda activate bandit_agent_dev_env

Exemples

I made a few exemples that can teach you the basics as to how to use the library

all exemples can be found in the Exemples folder of the project

Simulations

The simulations exemples are to help you develop an intuitive understanding of how the solvers perform depending on the context in which you put them

Basic simulations

Basic simulations are simply to show the performance of different solvers givent a very simple context

The basic simulations exemples are as follow

agents

Agents exemples are designed to show you the intended usage for the library

The agent exemples are as follow"

An actual exemple

Following is a simplified version of the basic agent exemple with more comment

Say we have two possible pseudo-random actions defined by func_a and func_b

def func_a(): float:
    return gamma.rvs(a=6.8, scale=0.1, loc=0, size=1)[0] # A mystery number

def func_b(): float:
    return gamma.rvs(a=2.2, scale=0.2, loc=0, size=1)[0] # A mystery number

for this exemple we will assume that we don't really know which of the action is better. We want to find out which action will return the best reward.

If we were to try and find that out by ourselves without any tool or help we'd have to execute the actions a few times and then remember which one tended to wield the best score after a few executions, then we would assume that said action is the "best" and exploit it by only calling it for a while. If after a while we doubt that the action we found is still best we would have to start that process all over again.

The purpose of Bandit Agents library is to automate that process, using different strategies that wield varying degrees of success given a few different contexts.

Let's continue our exemple; To exploit the best action we instanciate an agent with our mystery functions

action_a = ('action_a', func_a)
action_b = ('action_b', func_b)

actions = (action_a, action_b)

agent = Agent(actions)

We now have our agent, the agent is the object that will exploreand exploit our actions and will attemp to wield the best possible reward

Say we want the agent to work for a 100 epochs in which it will explore and exploit the actions we gave it, we would do that with the following code

for i in range(100):
    indexes, targets = agent.act()
    agent = agent.fit(x=indexes, y=targets)

If we want to see which action worked best, we can loog at the weights returned by the agent

print(agent.info())

Here is the complete example without interruption

from BanditAgents import Agent
from scipy.stats import gamma

def func_a(): float:
    return gamma.rvs(a=6.8, scale=0.1, loc=0, size=1)[0] # A mystery number

def func_b(): float:
    return gamma.rvs(a=2.2, scale=0.2, loc=0, size=1)[0] # A mystery number

action_a = ('action_a', func_a)
action_b = ('action_b', func_b)

actions = (action_a, action_b)

agent = Agent(actions)

for i in range(100):
    indexes, targets = agent.act()
    agent = agent.fit(x=indexes, y=targets)

print(agent.info())

Run exemples via the terminal

You can run all the exemples via the terminal. To do so go into the project's folder and run python3 -m . -e basic for the basic exemples and python3 -m . -e agents to run the agents exemples.

performances

To have get a breakdown of the agents performances I suggest you take a look at README.ipynb

There are many more options you can use in the library to solve your problems depending on your needs and I hope you will install it and use it for your projects.