MachineIntelligenceCore:ReinforcementLearning
 All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator
Class List
Here are the classes, structs, unions and interfaces with brief descriptions:
[detail level 123]
\Nmic
 oNalgorithms
 |\CMazeHistogramFilterClass implementing a histogram filter based solution of the maze-of-digits localization problem
 oNapplication
 |oCEpisodicHistogramFilterMazeLocalizationApplication for episodic testing of convergence of histogram filter based maze-of-digits localization
 |oCGridworldDeepQLearningClass responsible for solving the gridworld problem with Q-learning and (not that) deep neural networks
 |oCGridworldDRLExperienceReplayClass responsible for solving the gridworld problem with Q-learning, neural network used for approximation of the rewards and experience replay using for (batch) training of the neural network
 |oCGridworldDRLExperienceReplayPOMDPClass responsible for solving the gridworld problem with Q-learning, neural network used for approximation of the rewards and experience replay using for (batch) training of the neural network. In this case there is an assumption that the agent observes only part of the environment (POMPD)
 |oCGridworldQLearningClass responsible for solving the gridworld problem with Q-learning
 |oCGridworldValueIterationClass responsible for solving the gridworld problem by applying the reinforcement learning value iteration method
 |oCHistogramFilterMazeLocalizationClass implementing a histogram filter based solution of the maze-of-digits problem
 |oCMazeOfDigitsDLRERPOMPDApplication of Partially Observable Deep Q-learning with Experience Reply to the maze of digits problem. There is an assumption that the agent observes only part of the environment (POMPD)
 |oCMNISTDigitDLRERPOMDPApplication of Partially Observable Deep Q-learning with Experience Reply to the MNIST digits problem. There is an assumption that the agent observes only part of the environment - a patch of the whole image (POMPD)
 |oCTestAppClass implementing a n-Armed Bandits problem solving the n armed bandits problem using simple Q-learning rule
 |oCnArmedBanditsSofmaxClass implementing a n-Armed Bandits problem solving the n armed bandits problem using Softmax Action Selection
 |\CnArmedBanditsUnlimitedHistoryClass implementing a n-Armed Bandits problem solving the n armed bandits problem based on unlimited history action selection (storing all action-value pairs)
 oNenvironments
 |oCEnvironmentAbstract class representing an environment
 |oCGridworldClass emulating the gridworld environment
 |oCMazeOfDigitsClass emulating the maze of digits environment
 |\CMNISTDigitClass emulating the MNISTDigit digit environment
 oNimporters
 |\CMazeMatrixImporter
 \Ntypes
  oCSpatialExperienceStructure storing a spatial experience - a triplet of position in time t, executed action and position in time t+1
  \CSpatialExperienceMemoryClass representing the spatial experience memory - used in memory replay. Derived from the Batch class