Safer or Luckier? LLMs as Safety Evaluators Are Not Robust to ArtifactsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025 |
A review of faithfulness metrics for hallucination assessment in Large Language ModelsIEEE Journal on Selected Topics in Signal Processing (JSTSP), 2024 |
"Mango Mango, How to Let The Lettuce Dry Without A Spinner?'': Exploring
User Perceptions of Using An LLM-Based Conversational Assistant Toward
Cooking PartnerProceedings of the ACM on Human-Computer Interaction (PACMHCI), 2023 |