150 likes | 350 Views
A Reinforcement Learning Approach to Dynamic Resource Allocation. introduction. Dynamic resource allocation among multiple entities sharing a common set of resource The results of our predecessors (UP) Improvement (RL for U). Problem formulation.
E N D
A Reinforcement Learning Approach to Dynamic Resource Allocation
introduction • Dynamic resource allocation among multiple entities sharing a common set of resource • The results of our predecessors (UP) • Improvement (RL for U)
Problem formulation • Resource migrations require a non-negligible time • Algorithm for reassigning multiple resource units
Solution Methodology • Learning U in the dynamic resource allocation setting. • Predecessors:1.[8] single state of project. 2.[11] two values state but transfers of only a single resource type. • Improvement: extend [11] by considering transfers of multiple resource types. • dUi/dri=dUj/drj • n resource type ,s is n-dimensional vector • Rule base: advantage; disadvantage. DRA-FRL
Solution Methodology • Fuzzy Rulebase each parameter p gives the output value of the FRB when the input vector belongs to the categories A of rule i.
Solution Methodology • Reinforcement Learning Algorithm • A finite set of states S • A finite set of action A • A reward function r: S*A*S-----R • A state transition function T:S*A-----PD(S) • r(s,a,s)
Solution Methodology • Temporal difference (TD),TD(0)
Solution Methodology • Greed policy
Solution Methodology • [9]TDL with function approximation
Experimental Setup and Results • Queuing theory: M/D/n queue the expected queue length
Experimental Setup and Results • The optimal fixed resource allocation by queuing theory • Reactive policy balance resource utilization • Utility-based policy cost function:
Experimental Setup and Results • Step 1: use the “reactive” policy as the initial policy. • Step 2: DRA-FRL
Experimental Setup and Results • results