38
1
v1v2v3v4 (latest)

Learning distributed representations with efficient SoftMax normalization

Main:10 Pages
3 Figures
Bibliography:4 Pages
Appendix:5 Pages
Abstract

Learning distributed representations, or embeddings, that encode the relational similarity patterns among objects is a relevant task in machine learning. A popular method to learn the embedding matrices X,YX, Y is optimizing a loss function of the term SoftMax(XYT){\rm SoftMax}(XY^T). The complexity required to calculate this term, however, runs quadratically with the problem size, making it a computationally heavy solution. In this article, we propose a linear-time heuristic approximation to compute the normalization constants of SoftMax(XYT){\rm SoftMax}(XY^T) for embedding vectors with bounded norms. We show on some pre-trained embedding datasets that the proposed estimation method achieves higher or comparable accuracy with competing methods. From this result, we design an efficient and task-agnostic algorithm that learns the embeddings by optimizing the cross entropy between the softmax and a set of probability distributions given as inputs. The proposed algorithm is interpretable and easily adapted to arbitrary embedding problems. We consider a few use cases and observe similar or higher performances and a lower computational time than similar ``2Vec'' algorithms.

View on arXiv
@article{dallámico2025_2303.17475,
  title={ Learning distributed representations with efficient SoftMax normalization },
  author={ Lorenzo DallÁmico and Enrico Maria Belliardo },
  journal={arXiv preprint arXiv:2303.17475},
  year={ 2025 }
}
Comments on this paper