23
1

Representation Learning in a Decomposed Encoder Design for Bio-inspired Hebbian Learning

Abstract

Modern data-driven machine learning system designs exploit inductive biases in architectural structure, invariance and equivariance requirements, task-specific loss functions, and computational optimization tools. Previous works have illustrated that human-specified quasi-invariant filters can serve as a powerful inductive bias in the early layers of the encoder, enhancing robustness and transparency in learned classifiers. This paper explores this further within the context of representation learning with bio-inspired Hebbian learning rules. We propose a modular framework trained with a bio-inspired variant of contrastive predictive coding, comprising parallel encoders that leverage different invariant visual descriptors as inductive biases. We evaluate the representation learning capacity of our system in classification scenarios using diverse image datasets (GTSRB, STL10, CODEBRIM) and video datasets (UCF101). Our findings indicate that this form of inductive bias significantly improves the robustness of learned representations and narrows the performance gap between models using local Hebbian plasticity rules and those using backpropagation, while also achieving superior performance compared to non-decomposed encoders.

View on arXiv
@article{jaziri2025_2401.08603,
  title={ Representation Learning in a Decomposed Encoder Design for Bio-inspired Hebbian Learning },
  author={ Achref Jaziri and Sina Ditzel and Iuliia Pliushch and Visvanathan Ramesh },
  journal={arXiv preprint arXiv:2401.08603},
  year={ 2025 }
}
Comments on this paper