--- annotations_creators: - no-annotation language_creators: - crowdsourced language: - ace - ban - bjn - bug - gor - id - jv - mis - min - ms - nia - su - tet license: - cc-by-sa-3.0 - gfdl multilinguality: - multilingual source_datasets: - Wikipedia-HF task_categories: - text-generation - fill-mask task_ids: - language-modeling - masked-language-modeling pretty_name: Wikipedia Archive for Indonesian Languages & Local Languages tags: - Wikipedia - Untagged Languages ISO-639 (Banyumase/Ngapak) - Indonesian Language - Malaysian Language - Indonesia-related Languages - Indonesian Local Languages dataset_info: - config_name: indowiki_all features: - name: url dtype: string - name: title dtype: string - name: text dtype: string splits: - name: ace num_bytes: 4875688 num_examples: 12932 - name: ban num_bytes: 17561379 num_examples: 20243 - name: bjn num_bytes: 6669628 num_examples: 10460 - name: bug num_bytes: 3297641 num_examples: 15877 - name: gor num_bytes: 6007726 num_examples: 14572 - name: id num_bytes: 1103106307 num_examples: 657990 - name: jv num_bytes: 70335030 num_examples: 73150 - name: map_bms num_bytes: 5215803 num_examples: 13574 - name: min num_bytes: 116481049 num_examples: 227024 - name: ms num_bytes: 416001194 num_examples: 367463 - name: nia num_bytes: 1938378 num_examples: 1651 - name: su num_bytes: 47489084 num_examples: 61557 - name: tet num_bytes: 1452716 num_examples: 1465 download_size: 1803193334 dataset_size: 1800431623 - config_name: indowiki_dedup_all features: - name: url dtype: string - name: title dtype: string - name: text dtype: string splits: - name: ace num_bytes: 4867838 num_examples: 12904 - name: ban num_bytes: 17366080 num_examples: 19837 - name: bjn num_bytes: 6655378 num_examples: 10437 - name: bug num_bytes: 2072609 num_examples: 9793 - name: gor num_bytes: 5989252 num_examples: 14514 - name: id num_bytes: 1100932403 num_examples: 654287 - name: jv num_bytes: 69774853 num_examples: 72667 - name: map_bms num_bytes: 5060989 num_examples: 11832 - name: min num_bytes: 116376870 num_examples: 225858 - name: ms num_bytes: 410443550 num_examples: 346186 - name: nia num_bytes: 1938121 num_examples: 1650 - name: su num_bytes: 47410439 num_examples: 61494 - name: tet num_bytes: 1447926 num_examples: 1460 download_size: 1793103024 dataset_size: 1790336308 - config_name: indowiki_dedup_id_only features: - name: url dtype: string - name: title dtype: string - name: text dtype: string splits: - name: train num_bytes: 1100932403 num_examples: 654287 download_size: 1103131493 dataset_size: 1100932403 --- # **Indonesian Wikipedia Data Repository** --- license: cc-by-sa-3.0 --- Welcome to Indonesian Wikipedia Data Repository. The datasets are extracted from [Wikipedia HF](https://huggingface.co/datasets/wikipedia) and processed using the scripts available in this repository for reproducibility purpose. # **FAQS** ### How do I extract new Wikipedia Dataset of Indonesian languages? You may check to the script [_```extract_raw_wiki_data.py```_](https://huggingface.co/datasets/sabilmakbar/indonesian_wiki/blob/main/extract_raw_wiki_data.py) to understand its implementations, or you can adjust the bash provided in [_```extract_raw_wiki_data_indo.sh```_](https://huggingface.co/datasets/sabilmakbar/indonesian_wiki/blob/main/extract_raw_wiki_data_indo.sh) to extract it on your own. ### How do I extract new Wikipedia Dataset of Indonesian languages? You may visit this [Wikipedia Dump Index](https://dumps.wikimedia.org/backup-index.html) to check any latest available data and this link [Wikipedia Language Coverage](https://meta.wikimedia.org/wiki/List_of_Wikipedias#All_Wikipedias_ordered_by_number_of_articles) to map into any languages that you're wanting to extract. ### How does the data being preprocessed? What makes it different from loading it directly from Wikipedia HF? The data available in here are processed with following flows: 1. Raw data is being deduplicated on ```title``` and ```text``` (text-content from a given article), to remove articles containing boilerplate text (template text that are used usually for no-available informations or asking for contributions of content in that article), which usually deemed noisy for NLP data. 2. Furthermore, the ```title``` and ```text``` data are being checked for string-matching duplication (duplication of text that are being pre-processed, i.e symbols removed, HTML tags striped, or ASCII-chars validated). You may check this [ ```cleanse_wiki_data.py```](https://huggingface.co/datasets/sabilmakbar/indonesian_wiki/blob/main/cleanse_wiki_data.py) script to understand its implementation. # Getting Started # ### To read the datasets directly ### Use one of the following code chunks to load it from HuggingFace Hub: You can refer to the 2nd args of ```config name``` using the following script ``` dataset = load_dataset( "sabilmakbar/indonesian_wiki", "indo_wiki_dedup_data" # a config name, can be "indo_wiki_raw_data" or "indowiki_dedup_id_only", defaults to "indo_wiki_dedup_data" ) ``` Or you can provide both ```lang``` and ```date_stamp``` (providing only one will thrown an error) ``` dataset = load_dataset( "sabilmakbar/indonesian_wiki", lang = "id", # see the splits for complete lang choices date_stamp="20230901" ) ``` ### To replicate the whole dataset generation process ### 1. Set-up a new Python/Conda Environment (recommended Python version: 3.9.6 to 3.9.18 or 3.10.0 to 3.10.13) and install the requirements on ```requirements.txt``` use this codebase via ```pip install -r requirements.txt```. 2. Activate the chosen Python/Conda environment which the requirements are being installed. 3. Run this ```sh``` script for extractions from Wikimedia Dump: ```sh extract_raw_wiki_data_indo.sh```. 4. Run this ```sh``` script of deduplication: ```sh dedup_raw_wiki_data_indo.sh```. ## Citation Info: ``` @ONLINE{wikidump, author = "Wikimedia Foundation", title = "Wikimedia Downloads", url = "https://dumps.wikimedia.org"} @ONLINE{wikipedia-hf, title = "Huggingface Wikipedia Dataset", url = "https://huggingface.co/datasets/wikipedia"} ```