-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is there way to support yolo-obb model? #24
Comments
Good afternoon. |
Is there away to validate model output while using patch_inference as it usually boost the overall performance ? previously I was able to do so with a tool like sahi inference after transforming annotations to coco annotations. |
@Alisoltan82 Unfortunately, there isn't a ready-made tool to directly export patch inference results to COCO format. However, you can still obtain the results of the patch inference algorithm in a structured way, which you can then convert to any desired format, including COCO. Here's how you can access the results: # Final Results:
img = result.image
confidences = result.filtered_confidences
boxes = result.filtered_boxes
polygons = result.filtered_polygons
classes_ids = result.filtered_classes_id
classes_names = result.filtered_classes_names With these results, you can write the data into any format you need, including COCO. While we don't currently have a built-in tool for directly saving to COCO, we might consider adding such a feature in the future. For now, you can manually convert the results to COCO format using the data provided. |
Thanks |
Dear authors.
Do you have any plan to support yolo-obb model?
Thank you.
The text was updated successfully, but these errors were encountered: