Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's attention_mask to obtain reliable results. Traceback (most recent call last): #2

Open
yukePeng0815 opened this issue Dec 10, 2024 · 3 comments

Comments

@yukePeng0815
Copy link

image直接下载的您训练好的模型,运行测试代码报错

@taishan1994
Copy link
Owner

这个应该不是真正的错误,看下后面的截图。

@yukePeng0815
Copy link
Author

感谢已解决,device = "cuda:0" 加入这两行就好了
torch.backends.cuda.enable_mem_efficient_sdp(False)
torch.backends.cuda.enable_flash_sdp(False)

@yukePeng0815
Copy link
Author

您好想问一下,如果通过大模型实现信息抽取某个领域的三元组,微调需要的数据量需求最少是怎样的,如果不能微调可以通过prompt达到抽取的效果吗

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants