Dataset Viewer
	The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
			removing the
			loading script
			and relying on
			automated data support
			(you can use
			convert_to_parquet
			from the datasets library). If this is not possible, please
			open a discussion
			for direct help.
		
Dataset description
Number of articles for the English wiki:
- 2014: 4.599.592
- 2016: 5.144.403
- 2018: 5.599.764
- 2020: 6.037.287
- 2022: 6.291.973
- 2024: 6.629.861
with 8k documents per gzipped json, we get approx. 800 files per year, so aprrox. 3500 files in total. Approx. 190-200 MB per bin of 8k examples, 63 in gzipped mode.
Dataset sizes
There are different splits and dataset sizes, which are all subsets of the full set.
| Config | Train | Validation | Estimate | 
|---|---|---|---|
| tiny | 16k docs (2 shards) | 8k docs (1 shard) | 0.1GB | 
| small | 800k docs (100 shards) | 16k docs (2 shards) | 4GB | 
| medium | 6M docs (750 shards) | 16k docs (2 shards) | 30GB | 
| large | 12M docs (1500 shards) | 24k docs (3 shards) | 59GB | 
| full | 28M docs (3497 shards) | 32k docs (4 shards) | 137GB | 
- Downloads last month
- 15
