Files in MartinKies/USLR
Reinforcement Learning with R

DESCRIPTION
NAMESPACE
R/AsynchronousAdvantageActorCritic.R R/CuriosityFunctions.R R/DefaultFunctions.R R/HelpfullFunctions.R R/HotellingStratTourn.R R/NeuralNetwork.R R/PDstrategies.R R/PrisonersDilemmaStratTourn.R R/QlearningPersExpPath.R R/RFWager.R R/RNN.R R/SimpleGame.R R/ThesisFunctions.R R/XGBoost.R R/importPackages.R README.md Showcase Improved Q-Learning with Gradient Boosting.R Showcase Improved Q-Learning with RNN-LSTM.R Thesis Files/MainQLearningRNN_OptStratGeneratorThesis.R Thesis Files/MainQLearningXGB_OptStratGeneratorThesis.R man/Act.A3C.Rd man/Act.QLearningPersExpPath.Rd man/Action.2.Choice.PD.Rd man/Action.Encoding.Info.PD.Rd man/Action.Encoding.Info.Simple.Game.Rd man/Advantage.function.Rd man/Alphabet3.Rd man/Antimind.Rd man/Antimind2.Rd man/Antitiktak1.Rd man/Antitiktak1.improved.Rd man/Antitiktak1.simplified.Rd man/Antitiktak2.Rd man/Calc.R.phi.Rd man/Choice.2.Action.PD.Rd man/Choice.2.Action.Simple.Game.Rd man/Convert.2.train.Rd man/Define_Graph.Rd man/Define_Graph_Gradient_Update.Rd man/Discounted.Reward.PD.Rd man/Encode.Game.States.PD.Rd man/Encoding.Harper.PD.Rd man/Encoding.Manager.PD.Rd man/Encoding.last.X.rounds.PD.Rd man/Extend.Memory.QLearningPersExpPath.Rd man/External.Eval.PD.Rd man/Generate.Start.State.PD.Rd man/Generate.Start.State.Simple.Game.Rd man/Get.Def.Par.A3C.Rd man/Get.Def.Par.Neural.Network.A3C.LSTM.Rd man/Get.Def.Par.Neural.Network.A3C.Rd man/Get.Def.Par.Neural.Network.Rd man/Get.Def.Par.QLearningPersExpPath.Legacy.ThesisOpt.XGB.Rd man/Get.Def.Par.QLearningPersExpPath.Legacy.v.0.1.6.Rd man/Get.Def.Par.QLearningPersExpPath.Rd man/Get.Def.Par.RNN.Rd man/Get.Def.Par.XGBoost.Rd man/Get.Game.Object.PD.Rd man/Get.Game.Object.Simple.Game.Rd man/Get.Game.Param.PD.Rd man/Get.Par.PD.Rd man/Get.Par.Simple.Game.Rd man/Globaler.Tit.4.Tat.Rd man/Initialise.A3C.Rd man/Initialise.QLearningPersExpPath.Rd man/Memory.Random.Play.PD.Rd man/Memory.Self.Play.PD.Rd man/Mind.Rd man/MindHighD.Rd man/Model.strat.maximum.full.Ten.Rd man/NN.strat.Slim.TenTen.QLearning.Rd man/NN.strat.Slim.TenTen.Rd man/NN.strat.full.zero.Rd man/NN.strat.main.Rd man/NN.strat.static.end.Ten.Rd man/Play.Multiple.Games.QLearningPersExpPath.Rd man/Play.On.Strategy.QLearningPersExpPath.Rd man/Predict.Neural.Network.A3C.Rd man/Predict.Neural.Network.Rd man/Predict.RNN.Rd man/Q.on.hist.PD.QLearning.Rd man/Rainbow.Unicorns.one.Rd man/Rainbow.Unicorns.one.killer.Rd man/Rainbow.Unicorns.two.Rd man/Rainbow.Unicorns.two.killer.Rd man/Replay.QLearningPersExpPath.Rd man/Setup.Neural.Network.A3C.LSTM.Rd man/Setup.Neural.Network.A3C.Rd man/Setup.Neural.Network.Rd man/Setup.QLearningPersExpPath.Rd man/Setup.RNN.Rd man/State.2.Array.PD.Rd man/State.2.Array.Simple.Game.Rd man/State.Transition.PD.Rd man/State.Transition.Simple.Game.Rd man/TakTik1.Rd man/TakTik2.Rd man/TikTak1.Rd man/TikTak2.Rd man/TokTok.Rd man/Train.A3c.Rd man/Train.Neural.Network.Rd man/Train.On.Memory.QLearningPersExpPath.Rd man/Train.QLearningPersExpPath.Rd man/Train.RNN.Rd man/Update.Evaluator.QLearningPersExpPath.Rd man/Update.Memory.QLearningPersExpPath.Rd man/Weighted.Discount.Rd man/Worker.A3C.Rd man/a.Mind.Rd man/a.MindHighD.Rd man/a.TikTak1.Rd man/a.net.nice0.Rd man/a.ntft.1.Rd man/a.tadaaa.1.Rd man/a.td.2.Rd man/a.tiktak.2.Rd man/answer.strat1.Rd man/answer.strat2.Rd man/answer.strat3.Rd man/answer.to.undertaker.2.Rd man/antiprof2.Rd man/antistrat2.High.err.Rd man/antistrat2.Rd man/compare.exploration.Rd man/counter.Mind.Rd man/counter.MindHighD.Rd man/counter.da.taaa.Rd man/counter.nice.tit.for.tat.2.Rd man/counter.nice.tit.for.tat.Rd man/counter.rainbow.unicorns.two.Rd man/counter.strat2.Rd man/counter.ta.daaa.Rd man/counter.tft.forgive.fast.Rd man/counter.tft.forgive.slow.Rd man/counter.the.undertaker.1.Rd man/counter.the.undertaker.2.Rd man/da.taaa.Rd man/da.taaa.counter.Rd man/destab.strat2.0.Rd man/destab.ta.daaa.Rd man/dont.forgive.Rd man/dont.mind.High.err.Rd man/dont.mind.Rd man/eat.mindfreaks.2.Rd man/eat.unicorns.48.Rd man/eat.unicorns.50.Rd man/eat.unicorns.65.Rd man/el.majestro.incredibile.Rd man/false.friend.Rd man/fix.price.loc.Rd man/forgive.slower.Rd man/get.against.itself.benchmark.Rd man/get.antistrat.Rd man/get.benchmark.Rd man/get.conversion.Rd man/getrich.Rd man/harry.Rd man/into.spaaaace.Rd man/mean.tit.for.tat.Rd man/meineStrat2.Rd man/nashtag1.Rd man/net.nice.minus1.Rd man/net.nice.start1.Rd man/net.nice0.Rd man/net.nice1.Rd man/nice.tit.for.tat.High.err.Rd man/nice.tit.for.tat.Rd man/nottitfortat.Rd man/overtaker.High.err.Rd man/overtaker.Rd man/phases.Rd man/prep.data.4.shiny.Rd man/probably.nice.Base.Rd man/probably.nice.High.err.Rd man/probably.not.so.nice.Rd man/probably.not.that.nice.Rd man/prof.strat.Rd man/pudb.strat2.Rd man/redim.state.Rd man/regenbogen.Rd man/ron.Rd man/schachmatt_tournament.Rd man/screams.in.space.Rd man/seda.strat2.Rd man/smooth.average.Rd man/smooth.triangle.Rd man/squishy.the.octopus.Rd man/strat.fot.tik.tak.Rd man/strat1.Rd man/strat2.Rd man/strat2.killer.Rd man/strat3.Rd man/strat4.Rd man/stratego.Rd man/ta.daaa.Rd man/tatada.Rd man/tft.forgive.fast.2.Rd man/tft.forgive.fast.Rd man/tft.forgive.fast.killer.Rd man/tft.forgive.slow.Rd man/tft.forgive.slow.killer.Rd man/the.overtaker.1.Rd man/the.undertaker.1.Rd man/the.undertaker.2.Rd man/viva.PD.Strategy.Rd
MartinKies/USLR documentation built on Nov. 10, 2019, 5:24 a.m.