You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Upon investigation, I noticed that VisionPreProcessor is only initialized for mllama, llava_onevision, and qwen2_vl: Code Reference
However, 'llava' is included in an earlier assertion confirming it as a supported model type. This mismatch causes a failure when running inference.
Proposed Fix:
I recommend adding a llava_process method to VisionPreProcessor, ensuring LLaVA models correctly initialize preprocessing when needed: VisionPreProcessor class
Questions for Maintainers:
Was LLaVA deliberately excluded from the vision preprocessing logic?
Would extending VisionPreProcessor in this way be the best approach?
Are there other dependencies or configurations I should check before implementing this change?
Please advise on whether this approach aligns with your intended workflow. Thanks!
The text was updated successfully, but these errors were encountered:
While following the multimodal workflow guide for Triton Server, I encountered an assertion error:
Relevant Code
Upon investigation, I noticed that VisionPreProcessor is only initialized for
mllama
,llava_onevision
, andqwen2_vl
:Code Reference
However,
'llava'
is included in an earlier assertion confirming it as a supported model type. This mismatch causes a failure when running inference.Proposed Fix:
I recommend adding a
llava_process
method toVisionPreProcessor
, ensuring LLaVA models correctly initialize preprocessing when needed:VisionPreProcessor class
Questions for Maintainers:
VisionPreProcessor
in this way be the best approach?Please advise on whether this approach aligns with your intended workflow. Thanks!
The text was updated successfully, but these errors were encountered: