Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it not supported by streaming output? #6

Open
CaoChensy opened this issue Sep 11, 2024 · 7 comments
Open

Is it not supported by streaming output? #6

CaoChensy opened this issue Sep 11, 2024 · 7 comments

Comments

@CaoChensy
Copy link

Is it not supported by streaming output?

@polanwang404
Copy link

demo.py 中的 use_vllm = True # set True to use vllm for inference
就可以了,记得安装 vllm

@CaoChensy
Copy link
Author

Thanks!

@leoterry-ulrica
Copy link

demo.py 中的 use_vllm = True # set True to use vllm for inference 就可以了,记得安装 vllm

不起作用

@Neo-Zhangjiajie
Copy link
Contributor

Since we need to post-process the LLM output to match the citation number such as [6-8] with the context sentences, it currently does not support stream output.

@leoterry-ulrica
Copy link

Since we need to post-process the LLM output to match the citation number such as [6-8] with the context sentences, it currently does not support stream output.

@zRzRzRzRzRzRzR 说的不支持。

@polanwang404
Copy link

polanwang404 commented Sep 12, 2024

demo.py 中的 use_vllm = True # set True to use vllm for inference 就可以了,记得安装 vllm

不起作用

But after I set use_vllm to True, it did run, and achieved the similar result as the demo video
image

@leoterry-ulrica
Copy link

demo.py 中的 use_vllm = True # set True to use vllm for inference 就可以了,记得安装 vllm

不起作用

But after I set use_vllm to True, it did run, and achieved the similar result as the demo video image

这个跟流式输出无关。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants