RePro: Training Language Models to Faithfully Recycle the Web for Pretraining Paper • 2510.10681 • Published Oct 12 • 5
FLAME-MoE: A Transparent End-to-End Research Platform for Mixture-of-Experts Language Models Paper • 2505.20225 • Published May 26 • 3
DeepResearchGym: A Free, Transparent, and Reproducible Evaluation Sandbox for Deep Research Paper • 2505.19253 • Published May 25 • 32
MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models Paper • 2406.06046 • Published Jun 10, 2024 • 1
Data-Efficient Pretraining with Group-Level Data Influence Modeling Paper • 2502.14709 • Published Feb 20 • 1
OmniEval: An Omnidirectional and Automatic RAG Evaluation Benchmark in Financial Domain Paper • 2412.13018 • Published Dec 17, 2024 • 41