Act.A3C | Determines which action the algorithm takes |
Action.2.Choice.PD | Action to Array for Prisoners Dilemma |
Action.Encoding.Info.PD | Get Info of Action Encoding |
Action.Encoding.Info.Simple.Game | Get Info of Action Encoding |
Act.QLearningPersExpPath | Determines which action to take |
Advantage.function | Calculates N-Step Returns or weighted Temporal Difference... |
Calc.R.phi | deprecated |
Choice.2.Action.PD | Array to Action for Prisoners Dilemma |
Choice.2.Action.Simple.Game | Array to Action for Simple Game |
compare.exploration | Strategy highlighting differences in Exploration states are... |
Convert.2.train | Converts stored Memory into arrays. |
counter.strat.a | A best answer strategy |
counter.strat.c | A bast answer to a testpool strategy |
counter.strat.d | A best answer strategy |
counter.strat.e | A best answer strategy |
counter.strat.f | A best answer strategy |
counter.strat.g | A best answer strategy |
counter.strat.h | A best answer strategy |
counter.strat.i | A best answer strat |
Define_Graph | Graph for Network Loss according to A3C. |
Define_Graph_Gradient_Update | Graph to update Network weights |
Discounted.Reward.PD | Update Score based on expected Value of reward |
Encode.Game.States.PD | Transforms List of Gamestates to std encoding form |
Encoding.Harper.PD | Encoding based on Reinforcement Learning Produces Dominant... |
Encoding.last.X.rounds.PD | Flexible Encoding Function which expects the following... |
Encoding.Manager.PD | Internal Function to make working with different encodings... |
Extend.Memory.QLearningPersExpPath | Extend Memory by specified experiences |
External.Eval.PD | Evaluate the current strategy using StratTourn |
fix.price.loc | Example srategy for the Hotelling game |
Generate.Start.State.PD | Generates Start State for Prisoners Dilemma Game |
Generate.Start.State.Simple.Game | Generates Start State for Simple Game |
get.against.itself.benchmark | Payoff of strategy against itself |
get.antistrat | Get vector of names of counter strategy |
get.benchmark | Payoff of best answer against the strategy |
get.conversion | Conversion factor |
Get.Def.Par.A3C | Get Default Parameters of A3C. |
Get.Def.Par.Neural.Network | Define default Parameters of the Neural Network Function |
Get.Def.Par.Neural.Network.A3C | Get Default Parameters of the Feed-Forward Neural Network for... |
Get.Def.Par.Neural.Network.A3C.LSTM | Get Default Parameters of the LSTM Neural Network for the A3C... |
Get.Def.Par.NN.Legacy.Thesis.Basic | Default Parameters Basic Neural Network as used for the... |
Get.Def.Par.NN.Legacy.v.0.1.6 | Default Parameters Neural Network v.0.1.6 |
Get.Def.Par.QLearningPersExpPath | Default Parameters for (improved) Q-Learning |
Get.Def.Par.QLearningPersExpPath.Legacy.ThesisOpt.RNN | Default Parameters QLearningPersExpPath Recurrent NN v.0.1.6 |
Get.Def.Par.QLearningPersExpPath.Legacy.ThesisOpt.XGB | Default Parameters QLearningPersExpPath of the thesis of... |
Get.Def.Par.QLearningPersExpPath.Legacy.v.0.1.6 | Default Parameters QLearningPersExpPath v.0.1.6 |
Get.Def.Par.QLearningPersExpPath.QLearning.Basic | Default Parameters normal Q-Learnung used in the thesis of... |
Get.Def.Par.RNN | Define default Parameters of the RNN Function |
Get.Def.Par.RNN.Legacy.ThesisOpt | Default Parameters Recurrent Neural Network Thesis Martin... |
Get.Def.Par.RNN.Legacy.v.0.1.6 | Default Parameters Recurrent Neural Network version 0.1.6 |
Get.Def.Par.XGBoost | Default Parameters for XGBoost |
Get.Def.Par.XGBoost.Legacy.ThesisOpt | Default Parameters Gradient Boosting Thesis Martin Kies |
Get.Def.Par.XGBoost.Legacy.v.0.1.6 | Default Parameters Gradient Boosting version 0.1.6 |
Get.Game.Object.PD | Get Game Object which fully defines Prisoners Dilemma. |
Get.Game.Object.Simple.Game | Get Game Object which fully defines simple game. |
Get.Game.Param.PD | Defines the game environment of the IPD |
Get.Game.Param.PD.Legacy.BattleOfStrategies2013.Baseline | Default Parameters Game - Rules of Seminar 2013 |
Get.Game.Param.PD.Legacy.BattleOfStrategies2019 | Default Parameters Game - Rules of Seminar 2019 |
Get.Par.PD | Defines model parameters for 'Prisoners Dilemma' |
Get.Par.Simple.Game | Defines model parameters for 'Simple Game' |
Initialise.A3C | Set changeable A3C Parameters. |
Initialise.QLearningPersExpPath | Set changeable model variables |
Memory.Random.Play.PD | Generate Memory where strategies play against a random... |
Memory.Self.Play.PD | Generate Memory where strategies play against themselves |
Model.strat.maximum.full.Ten | A strategy to be used after model has been trained |
net.nice0 | A prof strategy |
net.nice1 | A variant to net.nice0 |
net.nice.minus1 | A variant to net.nice0 |
net.nice.start1 | A variant to net.nice0 |
NN.strat.full.zero | A strategy to be used after model has been trained |
NN.strat.main | The actual strategy after model has been trained |
NN.strat.Slim.TenTen | A strategy to be used after model has been trained |
NN.strat.Slim.TenTen.QLearning | A strategy to be used after model has been trained |
NN.strat.static.end.Ten | A strategy to be used after model has been trained |
PID.controller | PID controller, which generates ouptput based on the error |
Play.Multiple.Games.QLearningPersExpPath | Train multiple games |
Play.On.Strategy.QLearningPersExpPath | Play the game based on strategy |
Predict.Neural.Network | Evaluate Neural Network |
Predict.Neural.Network.A3C | Predict Neural Network |
Predict.RNN | Evaluate Recurrent RNN |
prep.data.4.shiny | Prepare Worker Memory to visualize with shiny |
Q.on.hist.PD.QLearning | Q-values based on history of IPD |
redim.state | Change dimensionality of the state array. |
Replay.QLearningPersExpPath | Train model of Q learning |
Setup.Neural.Network | Setup a Neural Network |
Setup.Neural.Network.A3C | Setup a Feed-Forward Neural Network for the... |
Setup.Neural.Network.A3C.LSTM | Setup a Neural Network with an LSTM-Layer for the... |
Setup.QLearningPersExpPath | Sets up a model based on model parameters |
Setup.RNN | Setup a RNN |
smooth.average | Calculates a sensible moving average based on smoothing... |
smooth.triangle | Calculates a sensible moving average based on smoothing... |
State.2.Array.PD | State to Array for Prisoners Dilemma |
State.2.Array.Simple.Game | State to Array for Simple Game |
State.Transition.PD | Get next State of Prisoners Dilemma Game |
State.Transition.Simple.Game | Get next State of Simple Game |
strat.a | A testpool strategy |
strat.b | A testpool strategy |
strat.c | A testpool strategy |
strat.d | A testpool strategy |
strat.e | A testpool strategy |
strat.f | A testpool strategy |
strat.g | A testpool strategy |
strat.h | A testpool strategy |
strat.i | A testpool strategy |
Train.A3c | Use the A3C algorithm to train a model |
Train.Neural.Network | Train Neural Network |
Train.On.Memory.QLearningPersExpPath | Trains model based on memory |
Train.QLearningPersExpPath | Train a model based on Q-Learning |
Train.RNN | Train RNN |
Update.Evaluator.QLearningPersExpPath | Controlled Copying of Models |
Update.Memory.QLearningPersExpPath | Add historic Q-Values and Curiosity to memory |
Weighted.Discount | Calculates a weighted Mean |
Worker.A3C | Defines an Agent based on the A3C-Algorithm |
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.