37
0

Dependency-aware Maximum Likelihood Estimation for Active Learning

Abstract

Active learning aims to efficiently build a labeled training set by strategically selecting samples to query labels from annotators. In this sequential process, each sample acquisition influences subsequent selections, causing dependencies among samples in the labeled set. However, these dependencies are overlooked during the model parameter estimation stage when updating the model using Maximum Likelihood Estimation (MLE), a conventional method that assumes independent and identically distributed (i.i.d.) data. We propose Dependency-aware MLE (DMLE), which corrects MLE within the active learning framework by addressing sample dependencies typically neglected due to the i.i.d. assumption, ensuring consistency with active learning principles in the model parameter estimation process. This improved method achieves superior performance across multiple benchmark datasets, reaching higher performance in earlier cycles compared to conventional MLE. Specifically, we observe average accuracy improvements of 6\%, 8.6\%, and 10.5\% for k=1k=1, k=5k=5, and k=10k=10 respectively, after collecting the first 100 samples, where entropy is the acquisition function and kk is the query batch size acquired at every active learning cycle.

View on arXiv
@article{kalkanli2025_2503.05969,
  title={ Dependency-aware Maximum Likelihood Estimation for Active Learning },
  author={ Beyza Kalkanli and Tales Imbiriba and Stratis Ioannidis and Deniz Erdogmus and Jennifer Dy },
  journal={arXiv preprint arXiv:2503.05969},
  year={ 2025 }
}
Comments on this paper