abdulhade commited on
Commit
db9e4c0
·
verified ·
1 Parent(s): f3871a1

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +125 -0
README.md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ku
4
+ size_categories:
5
+ - 1B<n<10B
6
+ ---
7
+
8
+ # Dataset Card for KurCorpus 2B
9
+
10
+ This dataset card documents KurCorpus 2B, a multidialectal Kurdish text corpus intended for large-scale language modeling and downstream NLP research.
11
+
12
+ This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
13
+
14
+ ## Dataset Details
15
+
16
+ ### Dataset Description
17
+
18
+ KurCorpus 2B is a large-scale Kurdish corpus with more than 2 billion tokens spanning the three major dialects of Kurdish: Sorani (ckb), Badini/Kurmanji (kmr), and Hawrami/Gorani (hac). It is released to advance research on low-resource languages and to support training of multidialectal language models such as KurBERT 2B.
19
+
20
+ - **Curated by:** Karwan Mahdi Rawf, Abdullah Abdullah, Amanj Hussein, Haukar Mohammed
21
+ - **Funded by [optional]:** More Information Needed
22
+ - **Shared by [optional]:** University of Halabja
23
+ - **Language(s) (NLP):** Kurdish Sorani (ckb), Kurmanji/Badini (kmr), Hawrami/Gorani (hac)
24
+ - **License:** CC BY 4.0
25
+
26
+ ### Dataset Sources [optional]
27
+
28
+ - **Repository:** https://huggingface.co/datasets/abdulhade/Kurdishcorpus
29
+ - **Paper [optional]:** More Information Needed
30
+ - **Demo [optional]:** More Information Needed
31
+ - **External record:** Mendeley Data DOI 10.17632/fb5xhhn6m5.1
32
+
33
+ ## Uses
34
+
35
+ ### Direct Use
36
+
37
+ - Pretraining and finetuning of Kurdish language models
38
+ - Text classification, sentiment analysis, and topic modeling in Kurdish
39
+ - Token-level tasks such as NER and POS tagging after suitable labeling
40
+ - Cross-dialect adaptation and transfer learning studies
41
+ - Machine translation experiments involving Kurdish dialects
42
+ - Information extraction, knowledge graph construction, and QA pipelines
43
+
44
+ ### Out-of-Scope Use
45
+
46
+ - High-stakes decision making without human oversight, such as medical, legal, or safety-critical advice
47
+ - De-anonymization, profiling, or identification of individuals
48
+ - Generating abusive or harmful content or targeted harassment
49
+ - Any use that violates source site terms or applicable laws
50
+
51
+ ## Dataset Structure
52
+
53
+ The dataset is provided as large compressed archives or text shards suitable for large-scale training. Common layouts:
54
+ - One or more archives (e.g., `.7z` or `.rar`) containing normalized text files
55
+ - Alternatively, multiple `.txt` or `.txt.gz` shards for streaming
56
+
57
+ If consumed as text shards, each sample is a line of UTF-8 text with a single field:
58
+ - **text**: string
59
+
60
+ No official train/validation/test splits are provided. Users should create splits appropriate to their tasks and ensure no domain leakage.
61
+
62
+ ## Dataset Creation
63
+
64
+ ### Curation Rationale
65
+
66
+ Kurdish is historically under-resourced with substantial dialectal variation and orthographic inconsistency. KurCorpus 2B aims to provide a large, cleaned, and consistent text resource to enable robust Kurdish NLP, including dialect-aware modeling and cross-dialect transfer.
67
+
68
+ ### Source Data
69
+
70
+ Sources include publicly accessible Kurdish text from:
71
+ - News websites
72
+ - Social media platforms such as Facebook and Telegram
73
+ - Digitized literary texts and online publications
74
+
75
+ #### Data Collection and Processing
76
+
77
+ A dialect-aware normalization and cleaning pipeline was applied:
78
+ - Unicode normalization (NFKC) for consistent code points
79
+ - Orthographic correction for Arabic-script variants (e.g., ك -> ک, ي -> ی)
80
+ - Noise removal of emojis, URLs, emails, and HTML with placeholders like [URL], [EMAIL]
81
+ - Whitespace and punctuation normalization
82
+ - Dialect-aware stopword filtering to reduce noise while preserving structure
83
+ - Sentence boundary and token-level cleaning to improve modeling readiness
84
+
85
+ Tools included standard Python text-processing libraries and custom normalization scripts. Exact heuristics and dictionaries can be expanded in future releases.
86
+
87
+ #### Who are the source data producers?
88
+
89
+ Original content was produced by Kurdish news outlets, social media users, and authors of literary or online texts in Kurdish. No private or non-public sources were intentionally collected.
90
+
91
+ ### Annotations [optional]
92
+
93
+ No manual annotations are included in this release.
94
+
95
+ #### Annotation process
96
+
97
+ Not applicable.
98
+
99
+ #### Who are the annotators?
100
+
101
+ Not applicable.
102
+
103
+ #### Personal and Sensitive Information
104
+
105
+ The corpus aggregates public web text and may include names of public figures or references to events. Emails and URLs were replaced with placeholders where detected, but residual personal or sensitive content may remain. Users must comply with the license and with applicable data protection laws in their jurisdictions.
106
+
107
+ ## Bias, Risks, and Limitations
108
+
109
+ - **Domain bias:** Over-representation of news and social media may skew styles and topics.
110
+ - **Dialect balance:** While multidialectal, some dialects or sub-varieties may be underrepresented.
111
+ - **Orthography and script:** Orthographic ambiguity and mixed conventions can persist despite normalization.
112
+ - **Safety:** Content may include sensitive or offensive text typical of web corpora.
113
+
114
+ ### Recommendations
115
+
116
+ - Apply task-specific filtering, deduplication, and domain balancing where needed.
117
+ - Use human-in-the-loop review for safety-critical or production deployments.
118
+ - Report dialectal coverage and distribution when publishing results.
119
+ - For benchmarking, create clean, transparent splits and avoid overlap with evaluation sets.
120
+
121
+ ## Citation [optional]
122
+
123
+ If you use KurCorpus 2B, please cite the Mendeley Data record:
124
+
125
+ **BibTeX:**