Likelihood Variance as Text Importance for Resampling Texts to Map Language Models (EMNLP 2025 Findings)
Likelihood Variance as Text Importance for Resampling Texts to Map Language Models
Momose Oyama, Ryo Kishino, Hiroaki Yamagiwa, Hidetoshi Shimodaira
arXiv:2505.15428 | accepted to EMNLP 2025 Findings
With approximately half the number of unique texts, both LS and KL sampling achieve model map errors comparable to those of uniform sampling.
💨 Code (generate data): fig2_resampling_error.py
🥒 Data (plot-ready): data/fig2_resampling_error.pkl
📙 Notebook (visualize): figure2.ipynb
LS sampling is as robust as uniform sampling, but requires fewer texts.
Using only texts selected through LS sampling allows new models to be efficiently added to the map.
💨 Code (generate data): fig3a_mapvariance.py | fig3b_addnew.py
🥒 Data (plot-ready): data/fig3a_mapvariance.pkl | data/fig3b_addnew.pkl
📙 Notebook (visualize): figure3.ipynb
Using model coordinates from unique texts, we predict the average performance across six downstream tasks with ridge regression.
See code_for_prediction/ for details.
modeldata_1018.pklis shared with the one in modelmap/1000models.tsne_Q.pklcontains the t-SNE coordinates of the 1018 models. The procedure to compute them is described intsne_Q.py.- The data in
./data/uniq-idx-weight/summarizes the results of each resampling method. These can be reproduced by runninguniq_idx_weight.py. - The model map with sampling error is visualized in
figure1.ipynb.
@inproceedings{oyama-etal-2025-likelihood,
author = {Momose Oyama and Ryo Kishino and Hiroaki Yamagiwa and Hidetoshi Shimodaira},
title = {Likelihood Variance as Text Importance for Resampling Texts to Map Language Models},
booktitle = {Findings of the Association for Computational Linguistics: EMNLP 2025},
year = {2025}
}


