ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.09792
9
0

Interim Report on Human-Guided Adaptive Hyperparameter Optimization with Multi-Fidelity Sprints

14 May 2025
Michael Kamfonas
ArXivPDFHTML
Abstract

This case study applies a phased hyperparameter optimization process to compare multitask natural language model variants that utilize multiphase learning rate scheduling and optimizer parameter grouping. We employ short, Bayesian optimization sessions that leverage multi-fidelity, hyperparameter space pruning, progressive halving, and a degree of human guidance. We utilize the Optuna TPE sampler and Hyperband pruner, as well as the Scikit-Learn Gaussian process minimization. Initially, we use efficient low-fidelity sprints to prune the hyperparameter space. Subsequent sprints progressively increase their model fidelity and employ hyperband pruning for efficiency. A second aspect of our approach is using a meta-learner to tune threshold values to resolve classification probabilities during inference. We demonstrate our method on a collection of variants of the 2021 Joint Entity and Relation Extraction model proposed by Eberts and Ulges.

View on arXiv
@article{kamfonas2025_2505.09792,
  title={ Interim Report on Human-Guided Adaptive Hyperparameter Optimization with Multi-Fidelity Sprints },
  author={ Michael Kamfonas },
  journal={arXiv preprint arXiv:2505.09792},
  year={ 2025 }
}
Comments on this paper