You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository is the official codebase of our paper "MiniLongBench: The Low-cost Long Context Understanding Benchmark for Large Language Models".
@@ -17,7 +17,7 @@ Through empirical analysis of over 60 LLMs, MiniLongBench reduces the average ev
17
17
18
18
2025-05 - We released MiniLongBench dataset in [[Baidu Drive]](https://pan.baidu.com/s/1vUq3C5JR3ICo_g8_JXxJ0w?pwd=6erx)[[Google Drive]](https://drive.google.com/drive/folders/1Ps1_VoI1ExI1ZvVbBSCEKBuJGUlTeMmA?usp=sharing)[[Hugging Face]](https://huggingface.co/datasets/linggm/MiniLongBench). 👈🎉Please try it!
19
19
20
-
2025-05 - Our paper "MiniLongBench" has been accepted to **ACL'25 main track**! [[Paper]](https://huggingface.co/datasets/linggm/MiniLongBench) 👈🎉Please read it!
20
+
2025-05 - Our paper "MiniLongBench" has been accepted to **ACL'25 main track**! [[Paper]](https://arxiv.org/abs/2505.19959) 👈🎉Please read it!
21
21
22
22
23
23
## ⚙️ Environment Setup
@@ -78,13 +78,3 @@ There are two evaluation methods for MiniLongBench.
78
78
2. Directly calculate the scores of LLMs on MiniLongBench (`eval_directly.ipynb`).
0 commit comments