# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import random import sys import time from functools import partial import numpy as np import paddle from paddlenlp.data import Stack, Tuple, Pad, Vocab from paddlenlp.datasets import load_dataset, MapDataset from paddlenlp.metrics import DetectionF1, CorrectionF1 from paddlenlp.transformers import ErnieModel, ErnieTokenizer from paddlenlp.transformers import LinearDecayWithWarmup from paddlenlp.utils.log import logger sys.path.append('../..') from pycorrector.ernie_csc.model import ErnieForCSC from pycorrector.ernie_csc.utils import convert_example, create_dataloader, read_train_ds # yapf: disable parser = argparse.ArgumentParser() parser.add_argument("--model_name_or_path", type=str, default="ernie-1.0", choices=["ernie-1.0"], help="Pretraining model name or path") parser.add_argument("--max_seq_length", type=int, default=128, help="The maximum total input sequence length after SentencePiece tokenization.") parser.add_argument("--learning_rate", type=float, default=5e-5, help="Learning rate used to train.") parser.add_argument("--batch_size", default=8, type=int, help="Batch size per GPU/CPU for training.") parser.add_argument("--save_steps", type=int, default=1000, help="Save checkpoint every X updates steps.") parser.add_argument("--logging_steps", type=int, default=1, help="Log every X updates steps.") parser.add_argument("--output_dir", type=str, default='checkpoints/', help="Directory to save model checkpoint") parser.add_argument("--epochs", type=int, default=3, help="Number of epoches for training.") parser.add_argument("--device", type=str, default="gpu", choices=["cpu", "gpu"], help="Select cpu, gpu devices to train model.") parser.add_argument("--seed", type=int, default=1, help="Random seed for initialization.") parser.add_argument("--weight_decay", default=0.01, type=float, help="Weight decay if we apply some.") parser.add_argument("--warmup_proportion", default=0.1, type=float, help="Linear warmup proption over the training process.") parser.add_argument("--max_steps", default=-1, type=int, help="If > 0: set total number of training steps to perform. Override num_train_epochs.", ) parser.add_argument("--pinyin_vocab_file_path", type=str, default="pinyin_vocab.txt", help="pinyin vocab file path") parser.add_argument("--adam_epsilon", default=1e-8, type=float, help="Epsilon for Adam optimizer.") parser.add_argument("--ignore_label", default=-1, type=int, help="Ignore label for CrossEntropyLoss") parser.add_argument("--extra_train_ds_dir", default=None, type=str, help="The directory of extra train dataset.") # yapf: enable args = parser.parse_args() def set_seed(args): random.seed(args.seed) np.random.seed(args.seed) paddle.seed(args.seed) @paddle.no_grad() def evaluate(model, eval_data_loader): model.eval() det_metric = DetectionF1() corr_metric = CorrectionF1() for step, batch in enumerate(eval_data_loader, start=1): input_ids, token_type_ids, pinyin_ids, det_labels, corr_labels, length = batch # det_error_probs shape: [B, T, 2] # corr_logits shape: [B, T, V] det_error_probs, corr_logits = model(input_ids, pinyin_ids, token_type_ids) det_metric.update(det_error_probs, det_labels, length) corr_metric.update(det_error_probs, det_labels, corr_logits, corr_labels, length) det_f1, det_precision, det_recall = det_metric.accumulate() corr_f1, corr_precision, corr_recall = corr_metric.accumulate() logger.info("Sentence-Level Performance:") logger.info("Detection metric: F1={:.4f}, Recall={:.4f}, Precision={:.4f}". format(det_f1, det_recall, det_precision)) logger.info("Correction metric: F1={:.4f}, Recall={:.4f}, Precision={:.4f}". format(corr_f1, corr_recall, corr_precision)) model.train() return det_f1, corr_f1 def do_train(args): set_seed(args) paddle.set_device(args.device) if paddle.distributed.get_world_size() > 1: paddle.distributed.init_parallel_env() pinyin_vocab = Vocab.load_vocabulary( args.pinyin_vocab_file_path, unk_token='[UNK]', pad_token='[PAD]') tokenizer = ErnieTokenizer.from_pretrained(args.model_name_or_path) ernie = ErnieModel.from_pretrained(args.model_name_or_path) model = ErnieForCSC( ernie, pinyin_vocab_size=len(pinyin_vocab), pad_pinyin_id=pinyin_vocab[pinyin_vocab.pad_token]) train_ds, eval_ds = load_dataset('sighan-cn', splits=['train', 'dev']) # Extend current training dataset by providing extra training # datasets directory. The suffix of dataset file name in extra # dataset directory has to be ".txt". The data format of # dataset need to be a couple of senteces at every line, such as: # "城府宫员表示,这是过去三十六小时内第三期强烈的余震。\t政府官员表示,这是过去三十六小时内第三起强烈的余震。\n" if args.extra_train_ds_dir is not None and os.path.exists( args.extra_train_ds_dir): data = train_ds.data data_files = [ os.path.join(args.extra_train_ds_dir, data_file) for data_file in os.listdir(args.extra_train_ds_dir) if data_file.endswith(".txt") ] for data_file in data_files: ds = load_dataset( read_train_ds, data_path=data_file, splits=["train"], lazy=False) data += ds.data train_ds = MapDataset(data) det_loss_act = paddle.nn.CrossEntropyLoss( ignore_index=args.ignore_label, use_softmax=False) corr_loss_act = paddle.nn.CrossEntropyLoss( ignore_index=args.ignore_label, reduction='none') trans_func = partial( convert_example, tokenizer=tokenizer, pinyin_vocab=pinyin_vocab, max_seq_length=args.max_seq_length) batchify_fn = lambda samples, fn=Tuple( Pad(axis=0, pad_val=tokenizer.pad_token_id), # input Pad(axis=0, pad_val=tokenizer.pad_token_type_id), # segment Pad(axis=0, pad_val=pinyin_vocab.token_to_idx[pinyin_vocab.pad_token]), # pinyin Pad(axis=0, dtype="int64"), # detection label Pad(axis=0, dtype="int64"), # correction label Stack(axis=0, dtype="int64") # length ): [data for data in fn(samples)] train_data_loader = create_dataloader( train_ds, mode='train', batch_size=args.batch_size, batchify_fn=batchify_fn, trans_fn=trans_func) eval_data_loader = create_dataloader( eval_ds, mode='eval', batch_size=args.batch_size, batchify_fn=batchify_fn, trans_fn=trans_func) num_training_steps = args.max_steps if args.max_steps > 0 else len( train_data_loader) * args.num_epochs lr_scheduler = LinearDecayWithWarmup(args.learning_rate, num_training_steps, args.warmup_proportion) logger.info("Total training step: {}".format(num_training_steps)) # Generate parameter names needed to perform weight decay. # All bias and LayerNorm parameters are excluded. decay_params = [ p.name for n, p in model.named_parameters() if not any(nd in n for nd in ["bias", "norm"]) ] optimizer = paddle.optimizer.AdamW( learning_rate=lr_scheduler, epsilon=args.adam_epsilon, parameters=model.parameters(), weight_decay=args.weight_decay, apply_decay_param_fun=lambda x: x in decay_params) global_steps = 1 best_f1 = -1 tic_train = time.time() for epoch in range(args.num_epochs): for step, batch in enumerate(train_data_loader, start=1): input_ids, token_type_ids, pinyin_ids, det_labels, corr_labels, length = batch det_error_probs, corr_logits = model(input_ids, pinyin_ids, token_type_ids) # Chinese Spelling Correction has 2 tasks: detection task and correction task. # Detection task aims to detect whether each Chinese charater has spelling error. # Correction task aims to correct each potential wrong charater to right charater. # So we need to minimize detection loss and correction loss simultaneously. # See more loss design details on https://aclanthology.org/2021.findings-acl.198.pdf det_loss = det_loss_act(det_error_probs, det_labels) corr_loss = corr_loss_act( corr_logits, corr_labels) * det_error_probs.max(axis=-1) loss = (det_loss + corr_loss).mean() loss.backward() optimizer.step() lr_scheduler.step() optimizer.clear_grad() if global_steps % args.logging_steps == 0: logger.info( "global step %d, epoch: %d, batch: %d, loss: %f, speed: %.2f step/s" % (global_steps, epoch, step, loss, args.logging_steps / (time.time() - tic_train))) tic_train = time.time() if global_steps % args.save_steps == 0: if paddle.distributed.get_rank() == 0: logger.info("Eval:") det_f1, corr_f1 = evaluate(model, eval_data_loader) f1 = (det_f1 + corr_f1) / 2 model_file = "model_%d" % global_steps if f1 > best_f1: # save best model paddle.save(model.state_dict(), os.path.join(args.output_dir, "best_model.pdparams")) logger.info("Save best model at {} step.".format( global_steps)) best_f1 = f1 model_file = model_file + "_best" model_file = model_file + ".pdparams" paddle.save(model.state_dict(), os.path.join(args.output_dir, model_file)) logger.info("Save model at {} step.".format(global_steps)) if args.max_steps > 0 and global_steps >= args.max_steps: return global_steps += 1 if __name__ == "__main__": do_train(args)