MLPMemory
Collection
5 items
•
Updated
•
1
A fine-tuned version of GPT2-large on the WikiText-103 dataset.
| Model | Perplexity | Improvement |
|---|---|---|
| GPT2-large (baseline) | 15.80 | - |
| GPT2-large-Finetuned | 10.42 | -5.38 |
This model was released as part of the paper "MLP Memory: A Retriever-Pretrained Memory for Large Language Models".
For more information, see: https://github.com/Binn0/MLPMemory.
If you use this model, please cite:
@inproceedings{Wei2025MLPMA,
title={MLP Memory: A Retriever-Pretrained Memory for Large Language Models},
author={Rubin Wei and Jiaqi Cao and Jiarui Wang and Jushi Kai and Qipeng Guo and Bowen Zhou and Zhouhan Lin},
year={2025},
url={https://api.semanticscholar.org/CorpusID:281658735}
}
Base model
openai-community/gpt2-xl