我正在使用
AutoModelForCausalLM和AutoTokenizer来生成文本输出DialoGPT。
无论出于何种原因,即使使用 Huggingface 提供的示例,我也会收到此警告:
正在使用仅解码器架构,但检测到右填充!为了正确的生成结果,请
padding_side='left'在初始化分词器时进行设置。
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-medium")
model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-medium")
# Let's chat for 5 lines
for step in range(5):
# encode the new user input, add the eos_token and return a tensor in Pytorch
new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt')
# append the new user input tokens to the chat history
bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids …Run Code Online (Sandbox Code Playgroud) python machine-learning huggingface-transformers huggingface-tokenizers