DocumentCode :
2615674
Title :
Optimizing time warp simulation with reinforcement learning techniques
Author :
Wang, Jun ; Tropp, Carl
Author_Institution :
McGill Univ., Montreal
fYear :
2007
fDate :
9-12 Dec. 2007
Firstpage :
577
Lastpage :
584
Abstract :
Adaptive time warp protocols in the literature are usually based on a pre-defined analytic model of the system, expressed as a closed form function that maps system state to control parameter. The underlying assumption is that this model itself is optimal. In this paper we present a new approach that utilizes reinforcement learning techniques, also known as simulation-based dynamic programming. Instead of assuming an optimal control strategy, the very goal of reinforcement learning is to find the optimal strategy through simulation. A value function that captures the history of system feedbacks is used, and no prior knowledge of the system is required. Our reinforcement learning techniques were implemented in a distributed VLSI simulator with the objective of finding the optimal size of a bounded time window. Our experiments using two benchmark circuits indicated that it was successful in doing so.
Keywords :
dynamic programming; learning (artificial intelligence); time warp simulation; adaptive time warp protocol; distributed VLSI simulator; reinforcement learning; simulation-based dynamic programming; system feedback; time warp simulation; Adaptive control; Control system synthesis; Dynamic programming; Feedback; History; Learning; Optimal control; Programmable control; Protocols; Time warp simulation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Simulation Conference, 2007 Winter
Conference_Location :
Washington, DC
Print_ISBN :
978-1-4244-1306-5
Electronic_ISBN :
978-1-4244-1306-5
Type :
conf
DOI :
10.1109/WSC.2007.4419650
Filename :
4419650
Link To Document :
بازگشت