nielsr HF Staff commited on
Commit
f224e8a
·
verified ·
1 Parent(s): 4fb56ef

Improve model card: Add Transformers library tag, paper abstract, and HF paper link

Browse files

This PR enhances the model card for `internlm/Intern-S1-mini-FP8` by:

- Adding `library_name: transformers` to the metadata, which enables the automated "how to use" widget on the model page and enhances discoverability for users.
- Updating the main title to include the paper name (`# Intern-S1-mini: A Scientific Multimodal Foundation Model`) for better clarity.
- Adding a direct link to the official Hugging Face paper page (`https://huggingface.co/papers/2508.15763`) for easy access to the canonical paper entry.
- Including the full paper abstract to provide a more comprehensive overview of the model's scientific contributions and capabilities.

Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -1,12 +1,13 @@
1
  ---
2
- license: apache-2.0
3
- pipeline_tag: image-text-to-text
4
  base_model:
5
  - internlm/Intern-S1-mini
 
 
 
6
  ---
7
 
8
-
9
- ## Intern-S1-mini
10
 
11
  <div align="center">
12
  <img src="https://cdn-uploads.huggingface.co/production/uploads/642695e5274e7ad464c8a5ba/E43cgEXBRWjVJlU_-hdh6.png" />
@@ -21,7 +22,9 @@ base_model:
21
  👋 join us on <a href="https://discord.gg/xa29JuW87d" target="_blank">Discord</a> and <a href="https://cdn.vansin.top/intern-s1.jpg" target="_blank">WeChat</a>
22
  </p>
23
 
 
24
 
 
25
 
26
  ## Introduction
27
 
@@ -464,4 +467,4 @@ If you find this work useful, feel free to give us a cite.
464
  primaryClass={cs.LG},
465
  url={https://arxiv.org/abs/2508.15763},
466
  }
467
- ```
 
1
  ---
 
 
2
  base_model:
3
  - internlm/Intern-S1-mini
4
+ license: apache-2.0
5
+ pipeline_tag: image-text-to-text
6
+ library_name: transformers
7
  ---
8
 
9
+ # Intern-S1-mini: A Scientific Multimodal Foundation Model
10
+ [📚 Hugging Face Paper](https://huggingface.co/papers/2508.15763)
11
 
12
  <div align="center">
13
  <img src="https://cdn-uploads.huggingface.co/production/uploads/642695e5274e7ad464c8a5ba/E43cgEXBRWjVJlU_-hdh6.png" />
 
22
  👋 join us on <a href="https://discord.gg/xa29JuW87d" target="_blank">Discord</a> and <a href="https://cdn.vansin.top/intern-s1.jpg" target="_blank">WeChat</a>
23
  </p>
24
 
25
+ ## Abstract
26
 
27
+ In recent years, a plethora of open-source foundation models have emerged, achieving remarkable progress in some widely attended fields, with performance being quite close to that of closed-source models. However, in high-value but more challenging scientific professional fields, either the fields still rely on expert models, or the progress of general foundation models lags significantly compared to those in popular areas, far from sufficient for transforming scientific research and leaving substantial gap between open-source models and closed-source models in these scientific domains. To mitigate this gap and explore a step further toward Artificial General Intelligence (AGI), we introduce Intern-S1, a specialized generalist equipped with general understanding and reasoning capabilities with expertise to analyze multiple science modal data. Intern-S1 is a multimodal Mixture-of-Experts (MoE) model with 28 billion activated parameters and 241 billion total parameters, continually pre-trained on 5T tokens, including over 2.5T tokens from scientific domains. In the post-training stage, Intern-S1 undergoes offline and then online reinforcement learning (RL) in InternBootCamp, where we propose Mixture-of-Rewards (MoR) to synergize the RL training on more than 1000 tasks simultaneously. Through integrated innovations in algorithms, data, and training systems, Intern-S1 achieved top-tier performance in online RL training. On comprehensive evaluation benchmarks, Intern-S1 demonstrates competitive performance on general reasoning tasks among open-source models and significantly outperforms open-source models in scientific domains, surpassing closed-source state-of-the-art models in professional tasks, such as molecular synthesis planning, reaction condition prediction, predicting thermodynamic stabilities for crystals. Our models are available at this https URL .
28
 
29
  ## Introduction
30
 
 
467
  primaryClass={cs.LG},
468
  url={https://arxiv.org/abs/2508.15763},
469
  }
470
+ ```