Datasets:
Tasks:
Text Generation
Formats:
parquet
Sub-tasks:
language-modeling
Languages:
Danish
Size:
10M - 100M
ArXiv:
DOI:
License:
Kenneth Enevoldsen
commited on
Added lex.dk
Browse files- README.md +35 -83
- data/lexdk/create.py +78 -0
- data/lexdk/descriptive_stats.json +1 -0
- data/lexdk/lexdk.md +77 -0
- data/lexdk/lexdk.parquet +3 -0
- data/opensubtitles/opensubtitles.md +0 -43
- src/tests/readme_parsing.py +1 -1
README.md
CHANGED
|
@@ -5,6 +5,10 @@ configs:
|
|
| 5 |
data_files:
|
| 6 |
- split: train
|
| 7 |
path: 'data/*/*.parquet'
|
|
|
|
|
|
|
|
|
|
|
|
|
| 8 |
- config_name: opensubtitles
|
| 9 |
data_files:
|
| 10 |
- split: train
|
|
@@ -116,7 +120,8 @@ language_bcp47:
|
|
| 116 |
|
| 117 |
<!--
|
| 118 |
readme structure is inspired by:
|
| 119 |
-
https://github.com/huggingface/datasets/blob/main/templates/README_guide.md
|
|
|
|
| 120 |
|
| 121 |
|
| 122 |
# 🧨 Danish Dynaword
|
|
@@ -138,6 +143,7 @@ https://github.com/huggingface/datasets/blob/main/templates/README_guide.md -->
|
|
| 138 |
- [Languages:](#languages)
|
| 139 |
- [Dataset Structure](#dataset-structure)
|
| 140 |
- [Data Instances](#data-instances)
|
|
|
|
| 141 |
- [Data Splits](#data-splits)
|
| 142 |
- [Dataset Creation](#dataset-creation)
|
| 143 |
- [Curation Rationale](#curation-rationale)
|
|
@@ -151,12 +157,6 @@ https://github.com/huggingface/datasets/blob/main/templates/README_guide.md -->
|
|
| 151 |
|
| 152 |
## Dataset Description
|
| 153 |
|
| 154 |
-
|
| 155 |
-
|
| 156 |
-
|
| 157 |
-
|
| 158 |
-
|
| 159 |
-
|
| 160 |
<!-- START-DESC-STATS -->
|
| 161 |
- **Language**: dan, dansk, Danish
|
| 162 |
- **Number of samples**: 576.59K
|
|
@@ -165,11 +165,6 @@ https://github.com/huggingface/datasets/blob/main/templates/README_guide.md -->
|
|
| 165 |
<!-- END-DESC-STATS -->
|
| 166 |
|
| 167 |
|
| 168 |
-
|
| 169 |
-
|
| 170 |
-
|
| 171 |
-
|
| 172 |
-
|
| 173 |
### Dataset Summary
|
| 174 |
|
| 175 |
The Danish dynaword is a continually developed collection of Danish free-form text datasets from various domains. It is intended to be continually updated with new data sources. If you would like to contribute a dataset see the [contribute section](#contributing-to-the-dataset)
|
|
@@ -221,13 +216,6 @@ The dataset contains text from different sources which are thoroughly defined in
|
|
| 221 |
|
| 222 |
Each entry in the dataset consists of a single text with associated metadata
|
| 223 |
|
| 224 |
-
|
| 225 |
-
|
| 226 |
-
|
| 227 |
-
|
| 228 |
-
|
| 229 |
-
|
| 230 |
-
|
| 231 |
<!-- START-SAMPLE -->
|
| 232 |
```py
|
| 233 |
{
|
|
@@ -259,13 +247,6 @@ An entry in the dataset consists of the following fields:
|
|
| 259 |
- `metadata/*`: Potentially additional metadata
|
| 260 |
<!-- END-SAMPLE -->
|
| 261 |
|
| 262 |
-
|
| 263 |
-
|
| 264 |
-
|
| 265 |
-
|
| 266 |
-
|
| 267 |
-
|
| 268 |
-
|
| 269 |
### Data Splits
|
| 270 |
|
| 271 |
The entire corpus is provided in the `train` split.
|
|
@@ -284,52 +265,35 @@ This data generally contains no annotation besides the metadata attached to each
|
|
| 284 |
|
| 285 |
Below follows a brief overview of the sources in the corpus along with their individual license.
|
| 286 |
|
| 287 |
-
|
| 288 |
-
|
| 289 |
-
|
| 290 |
-
|
| 291 |
-
|
| 292 |
-
|
| 293 |
-
|
| 294 |
-
|
| 295 |
-
|
| 296 |
-
|
| 297 |
-
|
| 298 |
-
|
| 299 |
-
|
| 300 |
-
|
| 301 |
-
|
| 302 |
-
|
| 303 |
-
|
| 304 |
-
|
| 305 |
-
|
| 306 |
<!-- START-MAIN TABLE -->
|
| 307 |
-
| Source | Description | N. Tokens
|
| 308 |
-
|
| 309 |
-
| [
|
| 310 |
-
| [
|
| 311 |
-
| [
|
| 312 |
-
| [
|
| 313 |
-
| [
|
| 314 |
-
| [
|
| 315 |
-
| [
|
| 316 |
-
| [
|
| 317 |
-
| [
|
| 318 |
-
| [
|
| 319 |
-
| [
|
| 320 |
-
| [
|
| 321 |
-
| [
|
| 322 |
-
| [
|
| 323 |
-
| [
|
| 324 |
-
| [
|
| 325 |
-
| [
|
| 326 |
-
| [
|
| 327 |
-
| [
|
| 328 |
-
| [
|
| 329 |
-
| [
|
| 330 |
-
| [
|
| 331 |
-
|
|
| 332 |
-
|
|
|
|
|
|
|
| 333 |
[opensubtitles]: data/opensubtitles/opensubtitles.md
|
| 334 |
[retsinformationdk]: data/retsinformationdk/retsinformationdk.md
|
| 335 |
[ep]: data/ep/ep.md
|
|
@@ -362,18 +326,6 @@ Below follows a brief overview of the sources in the corpus along with their ind
|
|
| 362 |
<!-- END-MAIN TABLE -->
|
| 363 |
|
| 364 |
|
| 365 |
-
|
| 366 |
-
|
| 367 |
-
|
| 368 |
-
|
| 369 |
-
|
| 370 |
-
|
| 371 |
-
|
| 372 |
-
|
| 373 |
-
|
| 374 |
-
|
| 375 |
-
|
| 376 |
-
|
| 377 |
You can learn more about each dataset by pressing
|
| 378 |
|
| 379 |
<!-- ### Quality Control
|
|
|
|
| 5 |
data_files:
|
| 6 |
- split: train
|
| 7 |
path: 'data/*/*.parquet'
|
| 8 |
+
- config_name: lexdk
|
| 9 |
+
data_files:
|
| 10 |
+
- split: train
|
| 11 |
+
path: data/lexdk/*.parquet
|
| 12 |
- config_name: opensubtitles
|
| 13 |
data_files:
|
| 14 |
- split: train
|
|
|
|
| 120 |
|
| 121 |
<!--
|
| 122 |
readme structure is inspired by:
|
| 123 |
+
https://github.com/huggingface/datasets/blob/main/templates/README_guide.md
|
| 124 |
+
-->
|
| 125 |
|
| 126 |
|
| 127 |
# 🧨 Danish Dynaword
|
|
|
|
| 143 |
- [Languages:](#languages)
|
| 144 |
- [Dataset Structure](#dataset-structure)
|
| 145 |
- [Data Instances](#data-instances)
|
| 146 |
+
- [Data Fields](#data-fields)
|
| 147 |
- [Data Splits](#data-splits)
|
| 148 |
- [Dataset Creation](#dataset-creation)
|
| 149 |
- [Curation Rationale](#curation-rationale)
|
|
|
|
| 157 |
|
| 158 |
## Dataset Description
|
| 159 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 160 |
<!-- START-DESC-STATS -->
|
| 161 |
- **Language**: dan, dansk, Danish
|
| 162 |
- **Number of samples**: 576.59K
|
|
|
|
| 165 |
<!-- END-DESC-STATS -->
|
| 166 |
|
| 167 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 168 |
### Dataset Summary
|
| 169 |
|
| 170 |
The Danish dynaword is a continually developed collection of Danish free-form text datasets from various domains. It is intended to be continually updated with new data sources. If you would like to contribute a dataset see the [contribute section](#contributing-to-the-dataset)
|
|
|
|
| 216 |
|
| 217 |
Each entry in the dataset consists of a single text with associated metadata
|
| 218 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 219 |
<!-- START-SAMPLE -->
|
| 220 |
```py
|
| 221 |
{
|
|
|
|
| 247 |
- `metadata/*`: Potentially additional metadata
|
| 248 |
<!-- END-SAMPLE -->
|
| 249 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 250 |
### Data Splits
|
| 251 |
|
| 252 |
The entire corpus is provided in the `train` split.
|
|
|
|
| 265 |
|
| 266 |
Below follows a brief overview of the sources in the corpus along with their individual license.
|
| 267 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 268 |
<!-- START-MAIN TABLE -->
|
| 269 |
+
| Source | Description | N. Tokens | License |
|
| 270 |
+
| :------------------ | :--------------------------------------------------------------------------------------------------------------------------- | :-------- | :--------------------- |
|
| 271 |
+
| [lexdk] | Permissible use articles from [lex.dk](https://lex.dk) | 5.69M | [CC-BY-SA 4.0] |
|
| 272 |
+
| [opensubtitles] | Danish subsection of [OpenSubtitles](https://opus.nlpl.eu/OpenSubtitles/corpus/version/OpenSubtitles) | 271.89M | [CC-0] |
|
| 273 |
+
| [retsinformationdk] | [retsinformation.dk](https://www.retsinformation.dk) (legal-information.dk) the official legal information system of Denmark | 516.54M | [Danish Copyright Law] |
|
| 274 |
+
| [ep] | The Danish subsection of [Europarl](https://aclanthology.org/2005.mtsummit-papers.11/) | 100.89M | [CC-0] |
|
| 275 |
+
| [ft] | Records from all meetings of The Danish parliament (Folketinget) in the parliament hall | 114.09M | [CC-0] |
|
| 276 |
+
| [wikisource] | The Danish subsection of [Wikisource](https://en.wikisource.org/wiki/Main_Page) | 5.34M | [CC-0] |
|
| 277 |
+
| [spont] | Conversational samples collected as a part of research projects at Aarhus University | 1.56M | [CC-0] |
|
| 278 |
+
| [tv2r] | Contemporary Danish newswire articles published between 2010 and 2019 | 21.67M | [CC-BY-SA 4.0] |
|
| 279 |
+
| [adl] | Danish literature from 1700-2023 from the Archive for Danish Literature (ADL) | 58.49M | [CC-0] |
|
| 280 |
+
| [hest] | Samples from the Danish debate forum www.heste-nettet.dk | 389.33M | [CC-0] |
|
| 281 |
+
| [skat] | Skat is the Danish tax authority. This dataset contains content from its website skat.dk | 122.12M | [CC-0] |
|
| 282 |
+
| [dannet] | [DanNet](https://cst.ku.dk/projekter/dannet) is a Danish WordNet | 1.52M | [DanNet 1.0 License] |
|
| 283 |
+
| [retspraksis] | Case law or judical practice in Denmark derived from [Retspraksis](https://da.wikipedia.org/wiki/Retspraksis) | 57.08M | [CC-0] |
|
| 284 |
+
| [wikibooks] | The Danish Subsection of [Wikibooks](https://www.wikibooks.org) | 6.24M | [CC-0] |
|
| 285 |
+
| [jvj] | The works of the Danish author and poet, [Johannes V. Jensen](https://da.wikipedia.org/wiki/Johannes_V._Jensen) | 3.55M | [CC-BY-SA 4.0] |
|
| 286 |
+
| [gutenberg] | The Danish subsection from Project [Gutenberg](https://www.gutenberg.org) | 6.76M | [Gutenberg License] |
|
| 287 |
+
| [botxt] | The Bornholmsk Ordbog Dictionary Projec | 847.97K | [CC-0] |
|
| 288 |
+
| [depbank] | The Danish subsection of the [Universal Dependencies Treebank](https://github.com/UniversalDependencies/UD_Danish-DDT) | 185.45K | [CC-BY-SA 4.0] |
|
| 289 |
+
| [naat] | Danish speeches from 1930-2022 | 286.68K | [CC-0] |
|
| 290 |
+
| [synne] | Dataset collected from [synnejysk forening's website](https://www.synnejysk.dk), covering the Danish dialect sønderjysk | 52.51K | [CC-0] |
|
| 291 |
+
| [wiki] | The Danish subsection of [wikipeadia](https://en.wikipedia.org/wiki/Main_Page) | 122.00M | [CC-0] |
|
| 292 |
+
| [nordjyllandnews] | Articles from the Danish Newspaper [TV2 Nord](https://www.tv2nord.dk) | 37.91M | [CC-0] |
|
| 293 |
+
| [relig] | Danish religious text from the 1700-2022 | 1.24M | [CC-0] |
|
| 294 |
+
| **Total** | | 1.85B | |
|
| 295 |
+
|
| 296 |
+
[lexdk]: data/lexdk/lexdk.md
|
| 297 |
[opensubtitles]: data/opensubtitles/opensubtitles.md
|
| 298 |
[retsinformationdk]: data/retsinformationdk/retsinformationdk.md
|
| 299 |
[ep]: data/ep/ep.md
|
|
|
|
| 326 |
<!-- END-MAIN TABLE -->
|
| 327 |
|
| 328 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 329 |
You can learn more about each dataset by pressing
|
| 330 |
|
| 331 |
<!-- ### Quality Control
|
data/lexdk/create.py
ADDED
|
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""download lexdk from alexandrainst/lexdk-open"""
|
| 2 |
+
|
| 3 |
+
from datetime import datetime
|
| 4 |
+
from pathlib import Path
|
| 5 |
+
from typing import cast
|
| 6 |
+
|
| 7 |
+
import pandas as pd
|
| 8 |
+
from datasets import Dataset, load_dataset
|
| 9 |
+
|
| 10 |
+
column_order = [
|
| 11 |
+
"text",
|
| 12 |
+
"source",
|
| 13 |
+
"id",
|
| 14 |
+
"added",
|
| 15 |
+
"created",
|
| 16 |
+
"license",
|
| 17 |
+
"domain",
|
| 18 |
+
"metadata",
|
| 19 |
+
]
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
def convert_sample(example: dict) -> dict:
|
| 23 |
+
# from sample:
|
| 24 |
+
# {
|
| 25 |
+
# "url": "https://denstoredanske.lex.dk/Kullmanns_M%C3%B8lle",
|
| 26 |
+
# "title": "Kullmanns Mølle",
|
| 27 |
+
# "clarification": "",
|
| 28 |
+
# "authors": ["https://brugere.lex.dk/6929"],
|
| 29 |
+
# "date": "2021-01-20T13:23:20+01:00",
|
| 30 |
+
# "license": "fri anvendelse",
|
| 31 |
+
# "text": "Kullmanns Mølle er en mølle i Gudhjem, opkaldt efter Matts Kullmann, der byggede møllen i 1893 til sin søn, Christian Kullmann, se Gudhjem Mølle.",
|
| 32 |
+
# }
|
| 33 |
+
date = datetime.fromisoformat(example["date"])
|
| 34 |
+
text = f"{example["title"]}\n\npubliceret: {date}\n{example["text"]}"
|
| 35 |
+
|
| 36 |
+
new_example = dict(
|
| 37 |
+
text_new=text,
|
| 38 |
+
id=example["url"],
|
| 39 |
+
source="lexdk",
|
| 40 |
+
domain="Conversation",
|
| 41 |
+
license="cc-by-sa-4.0",
|
| 42 |
+
added="2025-01-04",
|
| 43 |
+
created=f"{date.date()}, {date.date()}",
|
| 44 |
+
metadata={"source-pretty": "Lex.dk"},
|
| 45 |
+
)
|
| 46 |
+
|
| 47 |
+
return new_example
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
def main():
|
| 51 |
+
ds = load_dataset("alexandrainst/lexdk-open", split="train")
|
| 52 |
+
ds = cast(Dataset, ds)
|
| 53 |
+
|
| 54 |
+
dates = [datetime.fromisoformat(date).date() for date in ds["date"]]
|
| 55 |
+
print(str(min(dates)), ",", str(max(dates))) # 2009-01-28, 2023-09-05
|
| 56 |
+
|
| 57 |
+
assert len(set(ds["url"])) == len(ds)
|
| 58 |
+
|
| 59 |
+
ds = ds.map(convert_sample, num_proc=4)
|
| 60 |
+
ds = ds.select_columns(column_order[1:] + ["text_new"])
|
| 61 |
+
ds = ds.rename_columns({"text_new": "text"})
|
| 62 |
+
# ensure order
|
| 63 |
+
ds = ds.select_columns(column_order)
|
| 64 |
+
|
| 65 |
+
df = ds.to_pandas()
|
| 66 |
+
df = cast(pd.DataFrame, df)
|
| 67 |
+
dedup_df = df.drop_duplicates(keep="first", subset=["text"])
|
| 68 |
+
print("N. duplicates: ", df.shape[0] - dedup_df.shape[0]) # 0
|
| 69 |
+
|
| 70 |
+
ds = ds.select(dedup_df.index)
|
| 71 |
+
assert len(set(ds["text"])) == len(ds)
|
| 72 |
+
|
| 73 |
+
save_path = Path(__file__).parent / "lexdk.parquet"
|
| 74 |
+
ds.to_parquet(save_path)
|
| 75 |
+
|
| 76 |
+
|
| 77 |
+
if __name__ == "__main__":
|
| 78 |
+
main()
|
data/lexdk/descriptive_stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"number_of_samples": 11887, "average_document_length": 1405.6435601918063, "number_of_tokens": 5688613, "language": "dan, dansk, Danish", "revision": "546c3b35e0e37fe1f9eff91da9f73e5672833489"}
|
data/lexdk/lexdk.md
ADDED
|
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
pretty_name: OpenSubtitles
|
| 3 |
+
language:
|
| 4 |
+
- da
|
| 5 |
+
license: cc-by-sa-4.0
|
| 6 |
+
license_name: CC-BY-SA 4.0
|
| 7 |
+
task_categories:
|
| 8 |
+
- text-generation
|
| 9 |
+
- fill-mask
|
| 10 |
+
task_ids:
|
| 11 |
+
- language-modeling
|
| 12 |
+
source_datasets:
|
| 13 |
+
- alexandrainst/lexdk-open
|
| 14 |
+
---
|
| 15 |
+
|
| 16 |
+
# Dataset Card for OpenSubtitles
|
| 17 |
+
|
| 18 |
+
<!-- START-SHORT DESCRIPTION -->
|
| 19 |
+
Permissible use articles from [lex.dk](https://lex.dk).
|
| 20 |
+
<!-- END-SHORT DESCRIPTION -->
|
| 21 |
+
|
| 22 |
+
Lex.dk is a Danish online encyclopedia platform providing access to reliable and authoritative knowledge on a wide range of topics. It is created and curated by experts, ensuring high-quality, accurate content. The platform serves as a central hub for general and specialized information in Danish, making it a valuable resource for education, research, and general learning.
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
## Dataset Description
|
| 28 |
+
|
| 29 |
+
<!-- START-DESC-STATS -->
|
| 30 |
+
- **Language**: dan, dansk, Danish
|
| 31 |
+
- **Number of samples**: 11.89K
|
| 32 |
+
- **Number of tokens (Llama 3)**: 5.69M
|
| 33 |
+
- **Average document length (characters)**: 1405.64
|
| 34 |
+
<!-- END-DESC-STATS -->
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
## Dataset Structure
|
| 38 |
+
An example from the dataset looks as follows.
|
| 39 |
+
|
| 40 |
+
<!-- START-SAMPLE -->
|
| 41 |
+
```py
|
| 42 |
+
{
|
| 43 |
+
"text": "Oluf Høst Museet\n\npubliceret: 2014-04-23 03:42:33+02:00\nOluf Høst Museet, kunstmuseum i Gudhjem, Bor[...]",
|
| 44 |
+
"source": "lexdk",
|
| 45 |
+
"id": "https://denstoredanske.lex.dk/Oluf_H%C3%B8st_Museet",
|
| 46 |
+
"added": "2025-01-04",
|
| 47 |
+
"created": "2014-04-23, 2014-04-23",
|
| 48 |
+
"license": "cc-by-sa-4.0",
|
| 49 |
+
"domain": "Conversation",
|
| 50 |
+
"metadata": {
|
| 51 |
+
"source-pretty": "Lex.dk"
|
| 52 |
+
}
|
| 53 |
+
}
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
### Data Fields
|
| 57 |
+
|
| 58 |
+
An entry in the dataset consists of the following fields:
|
| 59 |
+
|
| 60 |
+
- `text`(`str`): The content of the document.
|
| 61 |
+
- `source` (`str`): The source of the document (see [Source Data](#source-data)).
|
| 62 |
+
- `id` (`str`): An unique identifier for each document.
|
| 63 |
+
- `added` (`str`): An date for when the document was added to this collection.
|
| 64 |
+
- `created` (`str`): An date range for when the document was originally created.
|
| 65 |
+
- `license` (`str`): The license of the document. The licenses vary according to the source.
|
| 66 |
+
- `domain` (`str`): The domain of the source
|
| 67 |
+
- `metadata/source-pretty` (`str`): The long form version of the short-form source name
|
| 68 |
+
- `metadata/*`: Potentially additional metadata
|
| 69 |
+
<!-- END-SAMPLE -->
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
## Additional Information
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
### Citation Information
|
| 76 |
+
|
| 77 |
+
This dataset is derived from the publicly availabe dataset [alexandrainst/lexdk-open](https://huggingface.co/datasets/alexandrainst/lexdk-open).
|
data/lexdk/lexdk.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c4779881f575d6f612c8603ed4896f10ebc7293c59637fa8a0773ee4545fce3
|
| 3 |
+
size 10007743
|
data/opensubtitles/opensubtitles.md
CHANGED
|
@@ -22,18 +22,6 @@ Danish subsection of [OpenSubtitles](https://opus.nlpl.eu/OpenSubtitles/corpus/v
|
|
| 22 |
|
| 23 |
## Dataset Description
|
| 24 |
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
<!-- START-DESC-STATS -->
|
| 38 |
- **Language**: dan, dansk, Danish
|
| 39 |
- **Number of samples**: 29.82K
|
|
@@ -42,31 +30,9 @@ Danish subsection of [OpenSubtitles](https://opus.nlpl.eu/OpenSubtitles/corpus/v
|
|
| 42 |
<!-- END-DESC-STATS -->
|
| 43 |
|
| 44 |
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
## Dataset Structure
|
| 58 |
An example from the dataset looks as follows.
|
| 59 |
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
|
| 63 |
-
|
| 64 |
-
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
<!-- START-SAMPLE -->
|
| 71 |
```py
|
| 72 |
{
|
|
@@ -99,15 +65,6 @@ An entry in the dataset consists of the following fields:
|
|
| 99 |
<!-- END-SAMPLE -->
|
| 100 |
|
| 101 |
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
| 110 |
-
|
| 111 |
### Additional Processing
|
| 112 |
|
| 113 |
Due to copyright concern additional documents have been removed due to copyright concerns. These include:
|
|
|
|
| 22 |
|
| 23 |
## Dataset Description
|
| 24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
<!-- START-DESC-STATS -->
|
| 26 |
- **Language**: dan, dansk, Danish
|
| 27 |
- **Number of samples**: 29.82K
|
|
|
|
| 30 |
<!-- END-DESC-STATS -->
|
| 31 |
|
| 32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
## Dataset Structure
|
| 34 |
An example from the dataset looks as follows.
|
| 35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
<!-- START-SAMPLE -->
|
| 37 |
```py
|
| 38 |
{
|
|
|
|
| 65 |
<!-- END-SAMPLE -->
|
| 66 |
|
| 67 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
### Additional Processing
|
| 69 |
|
| 70 |
Due to copyright concern additional documents have been removed due to copyright concerns. These include:
|
src/tests/readme_parsing.py
CHANGED
|
@@ -41,4 +41,4 @@ def replace_tag(markdown: str, package: str, tag: str) -> str:
|
|
| 41 |
start_md, _, remainder = markdown.partition(tag_start)
|
| 42 |
_, _, end_md = remainder.partition(tag_end)
|
| 43 |
|
| 44 |
-
return f"{start_md}
|
|
|
|
| 41 |
start_md, _, remainder = markdown.partition(tag_start)
|
| 42 |
_, _, end_md = remainder.partition(tag_end)
|
| 43 |
|
| 44 |
+
return f"{start_md}{tag_start}\n{package.strip()}\n{tag_end}{end_md}"
|