首页 PaddleNLP 帖子详情
RobertaTokenizer从本地路径加载模型报错
收藏
快速回复
PaddleNLP 问答NLP 590 0
RobertaTokenizer从本地路径加载模型报错
收藏
快速回复
PaddleNLP 问答NLP 590 0
    tokenizer.save_pretrained(args.output_dir)
    
    # 加载保存的模型和tokenizer
    model = model_class.from_pretrained(args.output_dir)
    tokenizer = tokenizer_class.from_pretrained(args.output_dir)
    model.to(args.device)
报错的是这一句:model = model_class.from_pretrained(args.output_dir), 这里model_class我设置为了RobertaModel,tokenizer_class设置为RobertaTokenizer。 报错信息如下:
[2022-11-10 16:20:49,907] [    INFO] - tokenizer config file saved in model_files/chinese_model/mrc/tokenizer_config.json
[2022-11-10 16:20:49,908] [    INFO] - Special tokens file saved in model_files/chinese_model/mrc/special_tokens_map.json
Traceback (most recent call last):
  File "ote_model/mrc_paddle/main.py", line 413, in 
    main()
  File "ote_model/mrc_paddle/main.py", line 390, in main
    tokenizer = tokenizer_class.from_pretrained(args.output_dir)
  File "/home/liangqiao2022/miniconda3/envs/paddle_env/lib/python3.8/site-packages/paddlenlp/transformers/roberta/tokenizer.py", line 665, in from_pretrained
    return RobertaBPETokenizer.from_pretrained(
  File "/home/liangqiao2022/miniconda3/envs/paddle_env/lib/python3.8/site-packages/paddlenlp/transformers/tokenizer_utils_base.py", line 1638, in from_pretrained
    tokenizer = cls(*init_args, **init_kwargs)
  File "/home/liangqiao2022/miniconda3/envs/paddle_env/lib/python3.8/site-packages/paddlenlp/transformers/utils.py", line 159, in __impl__
    init_func(self, *args, **kwargs)
TypeError: __init__() missing 2 required positional arguments: 'vocab_file' and 'merges_file'
为什么会这样?按道理说from_pretrained()函数后面两个参数不要也行。求大佬解释
0
收藏
回复
需求/bug反馈?一键提issue告诉我们
发现bug?如果您知道修复办法,欢迎提pr直接参与建设飞桨~
在@后输入用户全名并按空格结束,可艾特全站任一用户