File size: 3,569 Bytes
1b2a818
 
9bf3335
 
 
 
 
 
 
 
 
 
 
 
1b2a818
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9bf3335
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
license: mit
language:
- ro
tags:
- legal
- finance
- biology
- chemistry
- medical
- code
- climate
size_categories:
- 100K<n<1M
---
# Romanian Wikipedia - Processed Dataset

This is a processed version of the Romanian Wikipedia subset from the [FineWiki dataset](https://huggingface.co/datasets/HuggingFaceFW/finewiki), optimized for language model training and analysis. The dataset has been filtered to essential fields and enriched with token counts using tiktoken's cl100k_base encoding.

## Dataset Overview

- **Total Pages**: 493,462
- **Total Size**: ~669 MB (compressed parquet)
- **Language**: Romanian (ro)
- **Total Tokens**: ~395 million tokens (cl100k_base encoding)
- **Source**: FineWiki - An updated and better extracted version of the wikimedia/Wikipedia dataset originally released in 2023

## Dataset Structure

### Data Instances

Example from the Romanian subset (values truncated for readability):

```json
{
  "id": "rowiki/3217840",
  "title": "Melba (film)",
  "url": "https://ro.wikipedia.org/wiki/Melba_(film)",
  "date_modified": "2023-08-15T10:22:31Z",
  "text": "# Melba (film)\nMelba este un film biografic muzical britanic din 1953 regizat de Lewis Milestone...",
  "token_count": 2145
}
```

### Data Fields

- **id** (string): dataset-unique identifier; format: `rowiki/<page_id>`
- **title** (string): article title
- **url** (string): canonical article URL
- **date_modified** (string): ISO-8601 timestamp of the last page revision
- **text** (string): cleaned, structured article text preserving headings, lists, code/pre blocks, tables and math. Has markdown formatting (headings, tables, lists)
- **token_count** (int64): number of tokens in the text field, calculated using tiktoken's cl100k_base encoding (used by GPT-4 and similar models)

## Tokenization

Token counts are computed using [tiktoken](https://github.com/openai/tiktoken) with the **cl100k_base** encoding, which is the same tokenizer used by:
- GPT-4
- GPT-3.5-turbo
- text-embedding-ada-002

This makes the dataset particularly useful for training or fine-tuning models compatible with OpenAI's tokenization scheme.

## Processing Details

This dataset was created from the original FineWiki Romanian subset by:
1. Filtering from 14 columns down to 6 essential fields
2. Computing token counts for each article's text using tiktoken (cl100k_base)
3. Processing in batches of 10,000 rows for efficient computation
4. Saving as compressed parquet files with snappy compression

## Files

- `000_00000_processed.parquet`: 249,533 articles (~327 MB)
- `000_00001_processed.parquet`: 243,929 articles (~342 MB)

## Use Cases

- Training or fine-tuning Romanian language models
- Token budget analysis and dataset planning
- Information retrieval and semantic search
- Question answering systems
- Text summarization and generation tasks

## Citation Information

```bibtex
@dataset{penedo2025finewiki,
  author    = {Guilherme Penedo},
  title     = {FineWiki},
  year      = {2025},
  publisher = {Hugging Face Datasets},
  url       = {https://huggingface.co/datasets/HuggingFaceFW/finewiki},
  urldate   = {2025-10-20},
  note      = {Source: Wikimedia Enterprise Snapshot API (https://api.enterprise.wikimedia.com/v2/snapshots). Text licensed under CC BY-SA 4.0 with attribution to Wikipedia contributors.}
}
```

## License

The text content is licensed under **CC BY-SA 4.0** with attribution to Wikipedia contributors, as per the original Wikipedia content license.

## Dataset Creator

Processed and uploaded by [Yxanul](https://huggingface.co/Yxanul)