To use the dummy environment, one may start by specifying the state and action dimensions. More...
Classes  
class  Action 
Implementation of continuous action. More...  
class  State 
Implementation of state of the dummy environment. More...  
Public Member Functions  
State  InitialSample () 
Dummy function to mimic initial sampling in an environment. More...  
bool  IsTerminal (const State &) const 
Dummy function to find terminal state. More...  
double  Sample (const State &, const Action &, State &) 
Dummy function to mimic sampling in an environment. More...  
Static Public Attributes  
static size_t  rewardSize 
Dimensionality of the reward vector. Number of rewards. More...  
To use the dummy environment, one may start by specifying the state and action dimensions.
Eg:
Now the ContinuousActionEnv class can be used as an EnvironmentType in RL methods just as any other mlpack's implementation of gym environments.
Definition at line 124 of file env_type.hpp.

inline 
Dummy function to mimic initial sampling in an environment.
Definition at line 196 of file env_type.hpp.
References DiscreteActionEnv::State::State().

inline 
Dummy function to find terminal state.
*  (state) The current state. 
Definition at line 203 of file env_type.hpp.
Dummy function to mimic sampling in an environment.
*  (state) The current state. 
*  (action) The current action. 
*  (nextState) The next state. 
Definition at line 186 of file env_type.hpp.

static 
Dimensionality of the reward vector. Number of rewards.
Definition at line 206 of file env_type.hpp.