In this work, we consider an inversion attack on the obfuscated input embeddings sent to a language model on a server, where the adversary has no access to the language model or the obfuscation mechanism and sees only the obfuscated embeddings along with the model's embedding table. We propose BeamClean, an inversion attack that jointly estimates the noise parameters and decodes token sequences by integrating a language-model prior. Against Laplacian and Gaussian obfuscation mechanisms, BeamClean always surpasses naive distance-based attacks. This work highlights the necessity for and robustness of more advanced learned, input-dependent methods.
View on arXiv@article{kale2025_2505.13758, title={ BeamClean: Language Aware Embedding Reconstruction }, author={ Kaan Kale and Kyle Mylonakis and Jay Roberts and Sidhartha Roy }, journal={arXiv preprint arXiv:2505.13758}, year={ 2025 } }