Please use this identifier to cite or link to this item:
http://hdl.handle.net/10603/564545
Title: | Hyperparameter tuning using adaptive bayesian contextual hyperband for time series forecasting |
Researcher: | Lakshmi priya, S |
Guide(s): | Suresh jaganathan |
Keywords: | adaptive bayesian Computer Science Computer Science Information Systems Engineering and Technology Hyperparameter series forecasting |
University: | Anna University |
Completed Date: | 2024 |
Abstract: | In Machine learning, hyperparameters are all those parameters the user explicitly defines to enhance a model s learning. Although they are defined outside the model, hyperparameters significantly impact the model s behavior and performance. Tuning these hyperparameters plays a significant role in improving the model performance. Any tuning algorithm aims to efficiently navigate the hyperparameter search space to find the optimal configuration in minimal time. Different tuning algorithms proposed in the recent decade have gained popularity due to increased computational power. Random, grid search, algorithms involving Bayesian rule, bandit-based solutions, and evolutionary algorithms are some of the popularly used tuning algorithms. The commonality in these algorithms is finding a good trade-off between exploration versus exploitation of the search space, given the limited budget for tuning. With the same goal, this work proposes a hyperparameter tuning algorithm called Bayesian Contextual Hyperband (BCHB), incorporating Bayesian and multi-armed bandit techniques. The algorithm divides the total budget, the number of iterations or time allotted to each hyperparameter configuration, into different brackets that start with more exploration of the search space and end with the exploitation of the selected best configurations. Each bracket is further divided into iterations that select and evaluate a fixed number of configurations on an objective function. True objective function evaluation is an expensive step, and hence in the proposed algorithm, it is substituted by a surrogate model of hyperparameter configurations and their corresponding validation losses obtained from the history of past evaluations. newline |
Pagination: | xviii,131p. |
URI: | http://hdl.handle.net/10603/564545 |
Appears in Departments: | Faculty of Information and Communication Engineering |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01_title.pdf | Attached File | 169.74 kB | Adobe PDF | View/Open |
02_prelim pages.pdf | 2.69 MB | Adobe PDF | View/Open | |
03_content.pdf | 294.06 kB | Adobe PDF | View/Open | |
04_abstract.pdf | 229.39 kB | Adobe PDF | View/Open | |
05_chapter 1.pdf | 664.97 kB | Adobe PDF | View/Open | |
06_chapter 2.pdf | 722.71 kB | Adobe PDF | View/Open | |
07_chapter 3.pdf | 1.26 MB | Adobe PDF | View/Open | |
08_chapter 4.pdf | 1.28 MB | Adobe PDF | View/Open | |
09_chapter 5.pdf | 1.77 MB | Adobe PDF | View/Open | |
10_annexures.pdf | 84.43 kB | Adobe PDF | View/Open | |
80_recommendation.pdf | 74.41 kB | Adobe PDF | View/Open |
Items in Shodhganga are licensed under Creative Commons Licence Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0).
Altmetric Badge: