Table of Links
-
Related Works
-
Methodology
4.1 Formulation of the DRL Problem
4.2 Instance-Aware Deep Reinforcement Learning for Efficient Index Selection
-
Experiments
5.2 RL Training and Application Phase
The RL Training and Application Phase of IA2 transitions from initial preprocessing to actively engaging with defined action spaces and state representations, marked by:
TD3-TD-SWAR Algorithm Application: Leveraging the action space and state representations crafted in the preprocessing phase, IA2 employs the TD3-TD-SWAR algorithm, as outlined in Algorithm 1. Unlike merely operating on preprocessed data, this approach integrates action space restrictions—accounting for existing index candidates and their masking history. Each tuning step recalibrates masking
possibilities for subsequent selections, embodying a strategy that adaptively masks actions irrelevant based on the current agent states.
Adaptation to Workloads: Designed for flexibility, IA2 applies learned strategies to a range of workloads, efficiently adapting to both familiar and unseen environments, demonstrating its capability to handle diverse operational scenarios.
Authors:
(1) Taiyi Wang, University of Cambridge, Cambridge, United Kingdom ([email protected]);
(2) Eiko Yoneki, University of Cambridge, Cambridge, United Kingdom ([email protected]).
This paper is
