Mapping 1,000+ Language Models via the Log-Likelihood Vector
Momose Oyama, Hiroaki Yamagiwa, Yusuke Takase, Hidetoshi Shimodaira
arXiv:2502.16173 | accepted to ACL 2025 main
-
models_1018.yaml
A list of 1,018 model names used in our research. -
model-data-1018.pkl
Collected metadata for the 1,018 models, includingmodel_type,model_size, and other model attributes.
Usage examples can be found in load_model-data.ipynb.
texts-10k-pile.jsonl
A JSONL file containing 10,000 text chunks from the Pile dataset. Each line in the file is a JSON object representing one chunk, with fields such astext,pile_set_name, and indexing metadata.
-
raw_log-likelihood_1018.pkl
Log-likelihood data for 1,018 language models calculated on the texts-10k-pile.jsonl dataset.
Usage examples can be found in load_log-likelihood.ipynb. -
clipped_log-likelihood_1018.pkl
Log-likelihood data for 1,018 language models calculated on the texts-10k-pile.jsonl dataset.
This data is derived by clipping the bottom 2% of the values from theraw_log-likelihood_1018.pkldata.
Usage examples can be found in load_log-likelihood.ipynb.
-
calculate-log-likelihood.ipynb
Jupyter notebook containing sample code for calculating log-likelihood. -
For code that predicts model performance from model coordinates (Section 5) and generates the LaTeX file listing 1,018 models (Appendix L), see code_for_Section_5_and_Appendix_L/.
-
For the code for the experiments on weight interpolation (Section 6.3 and Appendix J), see code_for_weight_interpolation.
