ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.19212
36
0

Continual Reinforcement Learning for HVAC Systems Control: Integrating Hypernetworks and Transfer Learning

24 March 2025
Gautham Udayakumar Bekal
Ahmed Ghareeb
Ashish Pujari
    AI4CE
ArXivPDFHTML
Abstract

Buildings with Heating, Ventilation, and Air Conditioning (HVAC) systems play a crucial role in ensuring indoor comfort and efficiency. While traditionally governed by physics-based models, the emergence of big data has enabled data-driven methods like Deep Reinforcement Learning (DRL). However, Reinforcement Learning (RL)-based techniques often suffer from sample inefficiency and limited generalization, especially across varying HVAC systems. We introduce a model-based reinforcement learning framework that uses a Hypernetwork to continuously learn environment dynamics across tasks with different action spaces. This enables efficient synthetic rollout generation and improved sample usage. Our approach demonstrates strong backward transfer in a continual learning setting after training on a second task, minimal fine-tuning on the first task allows rapid convergence within just 5 episodes and thus outperforming Model Free Reinforcement Learning (MFRL) and effectively mitigating catastrophic forgetting. These findings have significant implications for reducing energy consumption and operational costs in building management, thus supporting global sustainability goals.Keywords: Deep Reinforcement Learning, HVAC Systems Control, Hypernetworks, Transfer and Continual Learning, Catastrophic Forgetting

View on arXiv
@article{bekal2025_2503.19212,
  title={ Continual Reinforcement Learning for HVAC Systems Control: Integrating Hypernetworks and Transfer Learning },
  author={ Gautham Udayakumar Bekal and Ahmed Ghareeb and Ashish Pujari },
  journal={arXiv preprint arXiv:2503.19212},
  year={ 2025 }
}
Comments on this paper