You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Feature request: ability to run annotation models ahead of time for an entire dataset and then load these from storage while performing interactive annotation.
Justification: In a data annotation project, the users who perform annotation often lack powerful GPUs. Even if they do have them, it can be difficult to get them working properly due to dependancies issues, especially in a corporate environment with tight security restrictions. However there is always at least one person (a data scientist usually) who does have a powerful working GPU, who can spare the time to run the model(s) ahead of time before distributing work to the annotation team. A secondary benefit is the elimination of waiting for the models to run after advancing to a new image.
Solution? Option to run the auto-annotation functions in batch mode and cache the results to disk for use in future sessions. In particular this would benefit SAM.
I am happy to contribute as a beginner!
The text was updated successfully, but these errors were encountered:
KDeser
changed the title
Pre-compute
Pre-compute automatic labelling models
Dec 7, 2024
Hi @KDeser,
Thank you for suggesting. It would be a very good feature! :D
Please go ahead and implement it.
I think we should keep the computed embedding in a way that people can copy and share them easily. What do you think about it?
Does @mnmnk43434 's idea and implementation help in some way?
As the main developer (or are you a bot?) do you have any suggestions on where to begin? We should try to avoid instantiating the ONNX models and related dependencies. Will modifying LRUCache class so it contains a match for every filename be enough?
Feature request: ability to run annotation models ahead of time for an entire dataset and then load these from storage while performing interactive annotation.
Justification: In a data annotation project, the users who perform annotation often lack powerful GPUs. Even if they do have them, it can be difficult to get them working properly due to dependancies issues, especially in a corporate environment with tight security restrictions. However there is always at least one person (a data scientist usually) who does have a powerful working GPU, who can spare the time to run the model(s) ahead of time before distributing work to the annotation team. A secondary benefit is the elimination of waiting for the models to run after advancing to a new image.
Solution? Option to run the auto-annotation functions in batch mode and cache the results to disk for use in future sessions. In particular this would benefit SAM.
I am happy to contribute as a beginner!
The text was updated successfully, but these errors were encountered: