From post

Decision Making from Confidence Measurement on the Reward Growth using Supervised Learning - A Study Intended for Large-scale Video Games.

, , , , и . ICAART (2), стр. 264-271. SciTePress, (2016)978-989-758-172-4.

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

Imitative Learning for Online Planning in Microgrids., , , , и . DARE, том 9518 из Lecture Notes in Computer Science, стр. 1-15. Springer, (2015)A Gaussian mixture approach to model stochastic processes in power systems., , , , и . PSCC, стр. 1-7. IEEE, (2016)Complementarity Assessment of South Greenland Katabatic Flows and West Europe Wind Regimes., , , , , , , , и . CoRR, (2018)Inferring bounds on the performance of a control policy from a sample of trajectories., , , и . ADPRL, стр. 117-123. IEEE, (2009)Decision Making from Confidence Measurement on the Reward Growth using Supervised Learning - A Study Intended for Large-scale Video Games., , , , и . ICAART (2), стр. 264-271. SciTePress, (2016)978-989-758-172-4.How to Discount Deep Reinforcement Learning: Towards New Dynamic Strategies., , и . CoRR, (2015)Optimistic planning for belief-augmented Markov Decision Processes., , и . ADPRL, стр. 77-84. IEEE, (2013)Using approximate dynamic programming for estimating the revenues of a hydrogen-based high-capacity storage device., , и . ADPRL, стр. 1-8. IEEE, (2014)Contributions to Batch Mode Reinforcement Learning.. University of Liège, Belgium, (2011)Aggregating Optimistic Planning Trees for Solving Markov Decision Processes., , и . NIPS, стр. 2382-2390. (2013)