--- language: - af - ar - az - bn - cs - de - en - es - et - fa - fi - fr - ga - gl - gu - he - hi - hr - id - it - ja - ka - kk - km - ko - lt - lv - mk - ml - mn - mr - my - ne - nl - pl - ps - pt - ro - ru - si - sl - sv - ta - th - tr - uk - ur - vi - xh - zh pretty_name: MegaWika Citation Source Text configs: - config_name: af data_files: - split: train path: af_0.jsonl - config_name: ar data_files: - split: train path: ar_0.jsonl - config_name: az data_files: - split: train path: az_0.jsonl - config_name: bn data_files: - split: train path: bn_0.jsonl - config_name: cs data_files: - split: train path: cs_0.jsonl - config_name: de data_files: - split: train path: de_0.jsonl - config_name: en data_files: - split: train path: en*.jsonl - config_name: es data_files: - split: train path: es_0.jsonl - config_name: et data_files: - split: train path: et_0.jsonl - config_name: fa data_files: - split: train path: fa_0.jsonl - config_name: fi data_files: - split: train path: fi_0.jsonl - config_name: fr data_files: - split: train path: fr*.jsonl - config_name: ga data_files: - split: train path: ga_0.jsonl - config_name: gl data_files: - split: train path: gl_0.jsonl - config_name: gu data_files: - split: train path: gu_0.jsonl - config_name: he data_files: - split: train path: he_0.jsonl - config_name: hi data_files: - split: train path: hi_0.jsonl - config_name: hr data_files: - split: train path: hr_0.jsonl - config_name: id data_files: - split: train path: id_0.jsonl - config_name: it data_files: - split: train path: it_0.jsonl - config_name: ja data_files: - split: train path: ja_0.jsonl - config_name: ka data_files: - split: train path: ka_0.jsonl - config_name: kk data_files: - split: train path: kk_0.jsonl - config_name: km data_files: - split: train path: km_0.jsonl - config_name: ko data_files: - split: train path: ko_0.jsonl - config_name: lt data_files: - split: train path: lt_0.jsonl - config_name: lv data_files: - split: train path: lv_0.jsonl - config_name: mk data_files: - split: train path: mk_0.jsonl - config_name: ml data_files: - split: train path: ml_0.jsonl - config_name: mn data_files: - split: train path: mn_0.jsonl - config_name: mr data_files: - split: train path: mr_0.jsonl - config_name: my data_files: - split: train path: my_0.jsonl - config_name: ne data_files: - split: train path: ne_0.jsonl - config_name: nl data_files: - split: train path: nl_0.jsonl - config_name: pl data_files: - split: train path: pl_0.jsonl - config_name: ps data_files: - split: train path: ps_0.jsonl - config_name: pt data_files: - split: train path: pt_0.jsonl - config_name: ro data_files: - split: train path: ro_0.jsonl - config_name: ru data_files: - split: train path: ru*.jsonl - config_name: si data_files: - split: train path: si_0.jsonl - config_name: sl data_files: - split: train path: sl_0.jsonl - config_name: sv data_files: - split: train path: sv_0.jsonl - config_name: ta data_files: - split: train path: ta_0.jsonl - config_name: th data_files: - split: train path: th_0.jsonl - config_name: tr data_files: - split: train path: tr_0.jsonl - config_name: uk data_files: - split: train path: uk_0.jsonl - config_name: ur data_files: - split: train path: ur_0.jsonl - config_name: vi data_files: - split: train path: vi_0.jsonl - config_name: xh data_files: - split: train path: xh_0.jsonl - config_name: zh data_files: - split: train path: zh_0.jsonl --- # MegaWika Citation Source Text ## Dataset Access You can access the dataset using the Hugging Face `datasets` library: ```python from datasets import load_dataset dataset = load_dataset("hltcoe/megawika-citation-text-only") ``` ## Dataset Information MegaWika Citation Source Text is a large-scale, multilingual collection of Wikipedia citations and their source text. ### Key Features - Covers 51 languages - Contains full text of Wikipedia citations - Mostly text-only version (little HTML or other markup) ### Data Format Each entry in the dataset is a JSON object with the following structure: ```json { "title": "Title of the Wikipedia article the citation came from", "last_revision": "The last revision of the Wikipedia page (not the citation)", "url": "URL of the citation", "source_text": "The citation source text" } ``` ## Language Files The dataset is split into multiple files, one for each language. For languages with a large amount of data, the files are further split into parts. Here's a sample of the available files: - af.jsonl - ar.jsonl - az.jsonl - bn.jsonl - cs.jsonl - de.jsonl - en.split.1.jsonl, en.split.2.jsonl, ..., en.split.8.jsonl - es.jsonl - fr.split.1.jsonl, fr.split.2.jsonl, fr.split.3.jsonl, fr.split.4.jsonl - ... For a complete list of available files, please refer to the dataset's file structure on Hugging Face. ## Usage Examples Here is an example of how you might use this dataset: ```python from datasets import load_dataset dataset = load_dataset("hltcoe/megawika-citation-text-only", "en")["train"] texts = dataset["source_text"] ```