REAct: Rational Exponential Activation for Better Learning and Generalization in PINNs
Physics-Informed Neural Networks (PINNs) offer a promising approach to simulating physical systems. Still, their application is limited by optimization challenges, mainly due to the lack of activation functions that generalize well across several physical systems. Existing activation functions often lack such flexibility and generalization power. To address this issue, we introduce Rational Exponential Activation (REAct), a generalized form of tanh consisting of four learnable shape parameters. Experiments show that REAct outperforms many standard and benchmark activations, achieving an MSE three orders of magnitude lower than tanh on heat problems and generalizing well to finer grids and points beyond the training domain. It also excels at function approximation tasks and improves noise rejection in inverse problems, leading to more accurate parameter estimates across varying noise levels.
View on arXiv@article{mishra2025_2503.02267, title={ REAct: Rational Exponential Activation for Better Learning and Generalization in PINNs }, author={ Sourav Mishra and Shreya Hallikeri and Suresh Sundaram }, journal={arXiv preprint arXiv:2503.02267}, year={ 2025 } }