-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add qurator-spk/sbb_binarization #214
Conversation
Co-authored-by: Stefan Weil <sw@weilnetz.de>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Are there already models available for production? Then a rule install-models-sbb-binarize
to download one or several models would be nice.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I second @stweil's comment about packaging models. Or can we do this via package_data
and the upcoming resource discovery feat?
Pre-trained models are available from https://qurator-data.de/sbb_binarization/ I'll add an As for bundling with the tool, I'll have to finish that in core first but then yes, certainly, they could be placed in one of the defined locations and |
But the Python package on PyPI should not be polluted with such large models. We could use setuptools'
Couldn't we add an environment variable (say |
Yes but then I'd suggest to implement it in a processor-independent way, see OCR-D/spec#176 |
That's not the same thing, though. Yours is an environment variable for a specific parameter (which is entirely new AFAIK). My suggestion was to follow the example of |
I realize this is more specific but it would not be much harder to implement in core for all the processors to make use of. I can implement (=steal from cor-asv-ann ;-)) |
That would be great. In particular for derived Docker images :-) |
|
@jbarth-ubhd has an example of the result in OCR-D/ocrd-website#172 (comment)