-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
blip-2 build error #695
Comments
@KKwanhee we have a fix internally and will be synced to github main branch soon. |
@QiJune Thanks for your work. I have a question, does trtiton_server support blip2? Could you help me? |
@jessiewiswjc Blip2 is not supported in TensorRT-LLM Triton backend yet. |
@KKwanhee The issue should have been fixed on the latest main branch, please kindly check. Thanks very much. Closing. Please let us know if you have questions. |
@kaiyux Thanks for your replay. Is there a plan for this(trtion for multi-modal)? |
I was following the guide for 'blip-2' to build, but encountered the following error in the section below.
trtllm-build --checkpoint_dir=./opt/2.7B/trt_ckpt/fp16/1-gpu/
--max_batch_size 8
--use_gpt_attention_plugin float16
--use_gemm_plugin float16
--max_input_len 924
--max_output_len 100
--max_beam_width 5
--output_dir ../blip2/trt_engine/blip-2-opt-2.7b/fp16/1-gpu
--max_prompt_embedding_table_size 256 # 256 = 32 (query_token number) * 8 (max_batch_size)
I also tried it on the 'rel' branch, but the issue persisted. Could you help me?
The text was updated successfully, but these errors were encountered: