You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to pass an input image along with the prompt to CodeAgent.run. I see that there is an additional_args argument but when I pass the image as {"image": "path/to/image.png"}, the agent ends up loading the image via pytesseract to read the contents of the image instead of passing it to OpenAI/Anthropic directly. Is there any way that I can ensure that the image is passed along with the prompt so that the model can infer information from it instead of using external libraries to load the image when using the LiteLLM integration?
@DarshanDeshpande
As far as I know, there's still no support for VLMs yet. An issue has been created for this already (to be added as a possible feature).
Hello,
I want to pass an input image along with the prompt to
CodeAgent.run
. I see that there is anadditional_args
argument but when I pass the image as{"image": "path/to/image.png"}
, the agent ends up loading the image via pytesseract to read the contents of the image instead of passing it to OpenAI/Anthropic directly. Is there any way that I can ensure that the image is passed along with the prompt so that the model can infer information from it instead of using external libraries to load the image when using the LiteLLM integration?My code for reference:
The text was updated successfully, but these errors were encountered: