Jingmei commited on
Commit
b2110e7
·
verified ·
1 Parent(s): 0811241

Upload folder using huggingface_hub

Browse files
dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train", "test"]}
dataset_info.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "source": "WOAH_Animal_Disease",
3
+ "context_length": 4096,
4
+ "model_checkpoint": "meta-llama/Llama-3.1-8B",
5
+ "max_chars_per_text": 200000,
6
+ "original_samples": 1740,
7
+ "train_chunks": 18590,
8
+ "test_chunks": 2021,
9
+ "total_chunks": 20611,
10
+ "vocab_size": 128256,
11
+ "batch_size": 100,
12
+ "num_proc": 20,
13
+ "seed": 123,
14
+ "test_size": 0.1,
15
+ "texts_pre_chunked": 310,
16
+ "use_fast_tokenizer": false
17
+ }
test/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0124c005dc7d265918a7669c8d770ae5172624ed04a43a036b19b14c02ed9df6
3
+ size 41407704
test/dataset_info.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "json",
3
+ "citation": "",
4
+ "config_name": "default",
5
+ "dataset_name": "json",
6
+ "dataset_size": 291206571,
7
+ "description": "",
8
+ "download_checksums": {
9
+ "/home/jingmei/Pandemic_LLM/Data_collection_others/./clean/WOAH_Animal_Disease.jsonl": {
10
+ "num_bytes": 291347554,
11
+ "checksum": null
12
+ }
13
+ },
14
+ "download_size": 291347554,
15
+ "features": {
16
+ "input_ids": {
17
+ "feature": {
18
+ "dtype": "int32",
19
+ "_type": "Value"
20
+ },
21
+ "_type": "Sequence"
22
+ },
23
+ "attention_mask": {
24
+ "feature": {
25
+ "dtype": "int8",
26
+ "_type": "Value"
27
+ },
28
+ "_type": "Sequence"
29
+ }
30
+ },
31
+ "homepage": "",
32
+ "license": "",
33
+ "size_in_bytes": 582554125,
34
+ "splits": {
35
+ "train": {
36
+ "name": "train",
37
+ "num_bytes": 291206571,
38
+ "num_examples": 1740,
39
+ "dataset_name": "json"
40
+ }
41
+ },
42
+ "version": {
43
+ "version_str": "0.0.0",
44
+ "major": 0,
45
+ "minor": 0,
46
+ "patch": 0
47
+ }
48
+ }
test/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "08ca08137577e617",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
tokenizer_WOAH_Animal_Disease.log ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-07-15 00:38 - Processing WOAH_Animal_Disease dataset...
2
+ 2025-07-15 00:38 - Parameters:
3
+ 2025-07-15 00:38 - - Data dir: ./clean
4
+ 2025-07-15 00:38 - - Output dir: ./tokenizated/WOAH_Animal_Disease
5
+ 2025-07-15 00:38 - - Model: meta-llama/Llama-3.1-8B
6
+ 2025-07-15 00:38 - - Context length: 4096
7
+ 2025-07-15 00:38 - - Batch size: 100
8
+ 2025-07-15 00:38 - - Num processes: 20
9
+ 2025-07-15 00:38 - - Test size: 0.1
10
+ 2025-07-15 00:38 - - Max chars per text: 200000
11
+ 2025-07-15 00:38 - - Use fast tokenizer: False
12
+ 2025-07-15 00:38 - Loading tokenizer...
13
+ 2025-07-15 00:38 - Tokenizer loaded. Vocab size: 128256
14
+ 2025-07-15 00:38 - Loading WOAH_Animal_Disease dataset...
15
+ 2025-07-15 00:38 - Original columns: ['disease', 'url', 'title', 'text', 'source', 'word_count']
16
+ 2025-07-15 00:38 - Loaded 1740 samples from WOAH_Animal_Disease
17
+ 2025-07-15 00:38 - Text length statistics (chars):
18
+ 2025-07-15 00:38 - - Min chars: 633
19
+ 2025-07-15 00:38 - - Max chars: 3,208,970
20
+ 2025-07-15 00:38 - - Mean chars: 158,495.7
21
+ 2025-07-15 00:38 - - Texts needing pre-chunking: 310 (17.8%)
22
+ 2025-07-15 00:38 - - Longest text: 3,208,970 chars
23
+ 2025-07-15 00:38 - Pre-chunking texts based on character count...
24
+ 2025-07-15 00:38 - After pre-chunking: 2665 text segments
25
+ 2025-07-15 00:38 - After pre-chunking text length statistics:
26
+ 2025-07-15 00:38 - - Max chars: 200,000
27
+ 2025-07-15 00:38 - - Mean chars: 103,496.8
28
+ 2025-07-15 00:38 - Shuffling and splitting dataset...
29
+ 2025-07-15 00:38 - Split - Train: 2398, Test: 267
30
+ 2025-07-15 00:38 - Starting tokenization...
31
+ 2025-07-15 00:42 - Tokenization complete for WOAH_Animal_Disease
32
+ 2025-07-15 00:42 - - Train samples: 2398
33
+ 2025-07-15 00:42 - - Test samples: 267
34
+ 2025-07-15 00:42 - Starting chunking to 4096 tokens...
35
+ 2025-07-15 00:45 - Chunking complete for WOAH_Animal_Disease
36
+ 2025-07-15 00:45 - - Train chunks: 18590
37
+ 2025-07-15 00:45 - - Test chunks: 2021
38
+ 2025-07-15 00:45 - Saving WOAH_Animal_Disease dataset...
39
+ 2025-07-15 00:45 - ✓ Processing complete for WOAH_Animal_Disease!
40
+ 2025-07-15 00:45 - Dataset saved to: ./tokenizated/WOAH_Animal_Disease
41
+ 2025-07-15 00:45 - Final stats:
42
+ 2025-07-15 00:45 - - Train chunks: 18,590
43
+ 2025-07-15 00:45 - - Test chunks: 2,021
44
+ 2025-07-15 00:45 - - Total chunks: 20,611
train/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0198a6710656af39bf8403c85b16652a0cb1f7cd4781dcb69d09d7770b882e91
3
+ size 380878248
train/dataset_info.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "json",
3
+ "citation": "",
4
+ "config_name": "default",
5
+ "dataset_name": "json",
6
+ "dataset_size": 291206571,
7
+ "description": "",
8
+ "download_checksums": {
9
+ "/home/jingmei/Pandemic_LLM/Data_collection_others/./clean/WOAH_Animal_Disease.jsonl": {
10
+ "num_bytes": 291347554,
11
+ "checksum": null
12
+ }
13
+ },
14
+ "download_size": 291347554,
15
+ "features": {
16
+ "input_ids": {
17
+ "feature": {
18
+ "dtype": "int32",
19
+ "_type": "Value"
20
+ },
21
+ "_type": "Sequence"
22
+ },
23
+ "attention_mask": {
24
+ "feature": {
25
+ "dtype": "int8",
26
+ "_type": "Value"
27
+ },
28
+ "_type": "Sequence"
29
+ }
30
+ },
31
+ "homepage": "",
32
+ "license": "",
33
+ "size_in_bytes": 582554125,
34
+ "splits": {
35
+ "train": {
36
+ "name": "train",
37
+ "num_bytes": 291206571,
38
+ "num_examples": 1740,
39
+ "dataset_name": "json"
40
+ }
41
+ },
42
+ "version": {
43
+ "version_str": "0.0.0",
44
+ "major": 0,
45
+ "minor": 0,
46
+ "patch": 0
47
+ }
48
+ }
train/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "458f3c0109a4cf0f",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }