From Insight to Exploit: Leveraging LLM Collaboration for Adaptive Adversarial Text GenerationConference on Empirical Methods in Natural Language Processing (EMNLP), 2025 |
Training with Fewer Bits: Unlocking Edge LLMs Training with Stochastic RoundingConference on Empirical Methods in Natural Language Processing (EMNLP), 2025 |
Elastic Architecture Search for Efficient Language ModelsIEEE International Conference on Multimedia and Expo (ICME), 2025 |
AttnCache: Accelerating Self-Attention Inference for LLM Prefill via Attention CacheIACR Cryptology ePrint Archive (IACR ePrint), 2025 |