Papers
Topics
Authors
Recent
Search
2000 character limit reached

Experience-Enhanced Learning: One Size Still does not Fit All in Automatic Database

Published 21 Nov 2021 in cs.DB | (2111.10813v1)

Abstract: Recent years, the database committee has attempted to develop automatic database management systems. Although some researches show that the applying AI to data management is a significant and promising direction, there still exists many problems in implementing these techniques to real applications (long training time, various environments and unstable performance). In this paper, we discover that traditional rule based methods have the potential to solve the above problems. We propose three methodologies for improving learned methods, i.e. label collection for efficiently pre-training, knowledge base for model transfer and theoretical guarantee for stable performance. We implement our methodologies on two widely used learning approaches, deep learning and reinforcement learning. Firstly, the novel experience enhanced deep learning (EEDL) could achieve efficient training and stable performance. We evaluate EEDL with cardinality estimation, an essential database management. The experimental results on four real dataset [1] show that our EEDL could outperforms the general DL model [2]. Secondly, we design a novel experience-enhanced reinforcement learning (EERL), which could efficiently converge and has better performance than general RL models [3]. We test EERL with online index tuning task. The experiments on TPC-H shows that EERL could accelerate the convergence of agent and generate better solution that generalizes the reinforcement learning.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.