MachineIntelligenceCore:ReinforcementLearning
 All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator
Class Hierarchy

Go to the graphical class hierarchy

This inheritance list is sorted roughly, but not completely, alphabetically:
[detail level 123]
oCApplication
|\Cmic::application::GridworldValueIterationClass responsible for solving the gridworld problem by applying the reinforcement learning value iteration method
oCImporter
|\Cmic::importers::MazeMatrixImporter
oCmic::algorithms::MazeHistogramFilterClass implementing a histogram filter based solution of the maze-of-digits localization problem
oCOpenGLApplication
|oCmic::application::HistogramFilterMazeLocalizationClass implementing a histogram filter based solution of the maze-of-digits problem
|oCmic::application::nArmedBanditsSofmaxClass implementing a n-Armed Bandits problem solving the n armed bandits problem using Softmax Action Selection
|oCmic::application::nArmedBanditsUnlimitedHistoryClass implementing a n-Armed Bandits problem solving the n armed bandits problem based on unlimited history action selection (storing all action-value pairs)
|\Cmic::application::TestAppClass implementing a n-Armed Bandits problem solving the n armed bandits problem using simple Q-learning rule
oCOpenGLEpisodicApplication
|oCmic::application::EpisodicHistogramFilterMazeLocalizationApplication for episodic testing of convergence of histogram filter based maze-of-digits localization
|oCmic::application::GridworldDeepQLearningClass responsible for solving the gridworld problem with Q-learning and (not that) deep neural networks
|oCmic::application::GridworldDRLExperienceReplayClass responsible for solving the gridworld problem with Q-learning, neural network used for approximation of the rewards and experience replay using for (batch) training of the neural network
|oCmic::application::GridworldDRLExperienceReplayPOMDPClass responsible for solving the gridworld problem with Q-learning, neural network used for approximation of the rewards and experience replay using for (batch) training of the neural network. In this case there is an assumption that the agent observes only part of the environment (POMPD)
|oCmic::application::GridworldQLearningClass responsible for solving the gridworld problem with Q-learning
|oCmic::application::MazeOfDigitsDLRERPOMPDApplication of Partially Observable Deep Q-learning with Experience Reply to the maze of digits problem. There is an assumption that the agent observes only part of the environment (POMPD)
|\Cmic::application::MNISTDigitDLRERPOMDPApplication of Partially Observable Deep Q-learning with Experience Reply to the MNIST digits problem. There is an assumption that the agent observes only part of the environment - a patch of the whole image (POMPD)
oCPropertyTree
|\Cmic::environments::EnvironmentAbstract class representing an environment
| oCmic::environments::GridworldClass emulating the gridworld environment
| oCmic::environments::MazeOfDigitsClass emulating the maze of digits environment
| \Cmic::environments::MNISTDigitClass emulating the MNISTDigit digit environment
oCmic::types::SpatialExperienceStructure storing a spatial experience - a triplet of position in time t, executed action and position in time t+1
\CSpatialExperienceBatch
 \Cmic::types::SpatialExperienceMemoryClass representing the spatial experience memory - used in memory replay. Derived from the Batch class