Skip to content

Conversation

@zwkkk
Copy link

@zwkkk zwkkk commented Apr 3, 2023

在最近更新的版本中,
1 tokenize_dataset_rows.py存在[150004] not in list的bug。
2 infer.ipynb中存在mask bug。

@mymusise
Copy link
Owner

mymusise commented Apr 7, 2023

感谢PR~

晚点我会测试下,没问题的话会merge

return_attention_mask=False,
add_special_tokens=False)
input_ids = prompt_ids + target_ids + [config.eos_token_id]
input_ids = prompt_ids + [150001, 150004] + target_ids + [150005]
Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这块修改和感觉是不是等价的?

prompt_ids = tokenizer.encode(**, add_special_tokens=True) == prompt_ids = tokenizer.encode(**, add_special_tokens=False) + [150001, 150004]

Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

如果用魔法数字会不会不太好,我看官方又更新了下token_id 🤣

Copy link
Owner

@mymusise mymusise Apr 7, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

看了下官方的预处理代码,好像用它提供的build_inputs_with_special_tokens更好

prompt_ids = tokenizer.encode(, add_special_tokens=False)
target_ids = tokenizer.encode(, add_special_tokens=False)
tokenizer.build_inputs_with_special_tokens(prompt_ids, target_ids )

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants