-
Notifications
You must be signed in to change notification settings - Fork 1
/
preprocess.py
79 lines (66 loc) · 3.3 KB
/
preprocess.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
# Copyright 2021 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ============================================================================
'''
Ernie preprocess script.
'''
import os
import argparse
from src.dataset import create_finetune_dataset, create_mrc_dataset
def parse_args():
"""set and check parameters."""
parser = argparse.ArgumentParser(description="ernie preprocess")
parser.add_argument("--task_type", type=str, default="false",
choices=["chnsenticorp", "xnli", "dbqa"],
help="Eval task type, default is msra_ner")
parser.add_argument("--eval_data_shuffle", type=str, default="false", choices=["true", "false"],
help="Enable eval data shuffle, default is false")
parser.add_argument("--eval_batch_size", type=int, default=1, help="Eval batch size, default is 1")
parser.add_argument("--eval_data_file_path", type=str, default="",
help="Data path, it is better to use absolute path")
parser.add_argument('--result_path', type=str, default='./preprocess_Result/', help='result path')
args_opt = parser.parse_args()
if args_opt.eval_data_file_path == "":
raise ValueError("'eval_data_file_path' must be set when do evaluation task")
return args_opt
if __name__ == "__main__":
args = parse_args()
args.eval_batch_size = 1
ds = create_finetune_dataset(batch_size=args.eval_batch_size,
repeat_count=1,
data_file_path=args.eval_data_file_path,
do_shuffle=(args.eval_data_shuffle.lower() == "true"))
ids_path = os.path.join(args.result_path, "00_data")
mask_path = os.path.join(args.result_path, "01_data")
token_path = os.path.join(args.result_path, "02_data")
label_path = os.path.join(args.result_path, "03_data")
os.makedirs(ids_path)
os.makedirs(mask_path)
os.makedirs(token_path)
os.makedirs(label_path)
for idx, data in enumerate(ds.create_dict_iterator(output_numpy=True, num_epochs=1)):
input_ids = data["input_ids"]
input_mask = data["input_mask"]
token_type_id = data["token_type_id"]
label_ids = data["label_ids"]
file_name = args.task_type + "_batch_" + str(args.eval_batch_size) + "_" + str(idx) + ".bin"
ids_file_path = os.path.join(ids_path, file_name)
input_ids.tofile(ids_file_path)
mask_file_path = os.path.join(mask_path, file_name)
input_mask.tofile(mask_file_path)
token_file_path = os.path.join(token_path, file_name)
token_type_id.tofile(token_file_path)
label_file_path = os.path.join(label_path, file_name)
label_ids.tofile(label_file_path)
print("=" * 20, "export bin files finished", "=" * 20)