-
Notifications
You must be signed in to change notification settings - Fork 3.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] Tracks with AI agents #9054
Comments
This is correct at the moment. However, we're planning to expand agent capabilities to cover tracking as well. |
@SpecLad , thanks for the answer. I think we can close this? Or you'll close it when tracks are in place? |
@SpecLad , another question here ) |
I haven't considered it, though in principle it could be done. Could you explain your use case for a feature like this? |
Sure. Imagine a large image, where a car is visible. The task of extracting the car license plate from the full image is vastly simplified if the OCR model is given both the image data and information about the location of the plate (bbox) inside that image. For this reason, having the ability to send both image and extra data (such as bboxes) might be important. In the above scenario, the bbox of the license plate would be created in CVAT via the UI, or potentially by another model that detects license plates. |
Hello!
Recently we've found out this https://www.cvat.ai/blog/announcing-cvat-ai-agents#what-is-a-cvat-ai-agent-
As far as I understand the process, right now it is impossible to output tracks from any model which can be mounted to the agent due to frame-by-frame approach.
So, we feed the model with images frame by frame and get annotations from this model also for each separate frame => no tracks possible
Am i right?
Thanks!
The text was updated successfully, but these errors were encountered: