Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

当prompt_token 超过模型支持的最大长度时, 程序就恢复不了, 一直返回报错信息 #28

Open
yejunjin opened this issue Jul 3, 2024 · 2 comments

Comments

@yejunjin
Copy link
Collaborator

yejunjin commented Jul 3, 2024

在dashinfer集成进fastchat过程中,当prompt token超过engine_max_length时 当.generation_config.max_length < prompt token < .engine_config.engine_max_length,程序恢复不了。

@yejunjin
Copy link
Collaborator Author

yejunjin commented Jul 3, 2024

当config.json文件中,.engine_config.engine_max_length = 128, .generation_config.max_length = 64, 输入提问长度80左右,就会复现。

复现日志如下:
image

原因:主要是

runDecoderContext();
这段代码没有进行status判断,返回了错误状态也继续执行了。

@chuanzhubin
Copy link

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants