net.sourceforge.jabm.learning
Interface MDPLearner

All Superinterfaces:
DiscreteLearner, Learner
All Known Implementing Classes:
QLearner

public interface MDPLearner
extends DiscreteLearner

Classes implementing this interface implement learning algorithms for Markoff descision processes (MDPs).

 

Method Summary
 int bestAction(int state)
           
 int getNumberOfActions()
          Get the number of different possible actions this learner can choose from when it performs an action.
 int getNumberOfStates()
           
 double[] getValueEstimates(int state)
           
 void newState(double reward, int newState)
          The call-back after performing an action.
 
Methods inherited from interface net.sourceforge.jabm.learning.DiscreteLearner
act
 
Methods inherited from interface net.sourceforge.jabm.learning.Learner
dumpState, getLearningDelta, monitor
 

Method Detail

newState

void newState(double reward,
              int newState)
The call-back after performing an action.

Parameters:
reward - The reward received from taking the most recently-selected action.
newState - The new state encountered after taking the most recently-selected action.

getValueEstimates

double[] getValueEstimates(int state)
Parameters:
state - The current state of the MDP.
Returns:
An array representing the Q values indexed by action.

bestAction

int bestAction(int state)

getNumberOfActions

int getNumberOfActions()
Description copied from interface: DiscreteLearner
Get the number of different possible actions this learner can choose from when it performs an action.

Specified by:
getNumberOfActions in interface DiscreteLearner
Returns:
An integer value representing the number of actions available.

getNumberOfStates

int getNumberOfStates()