You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am actually working on training Grantha script. What I figured out is that both Tamil and Malayalam have almost same characters and formatting as that of Grantha.
So, I assume if both models are used simutaneously, then the combined model can be used for finetuning using my Grantha data. I am not sure whether it will work. What is your opinion? Is this the only way like finetuning each model separately and then use this command:
"tesseract image.png -l grantha+grantha1"
The text was updated successfully, but these errors were encountered:
Have you had any success in developing a model for Grantha script? There are others willing to collaborate on such a project. Please respond with any progress you have made. Thanks.
Indeed. I trained a model which is fine-tuned from Tamil.traineddata (from tessdata_best repository). But it would work better on printed images but mercilessly fail on handwritten images due to segmentation issues.
Still there are exceptions. I can train handwritten images with single line up to 4 lines. But after that I am getting the error "Compute CTC targets failed for ....". And I raised it as an issue in #414 .
After delving into this training approach, I came up with a conclusion. I assume some upgradation would be required both in the VGSL Network Specification to include more advanced network layers, and in PSM definition for recognizing handwritten images apart from LSTM networks.
It would be grateful if I could be a part of that project. This would really help me elevate my career.
I am actually working on training Grantha script. What I figured out is that both Tamil and Malayalam have almost same characters and formatting as that of Grantha.
So, I assume if both models are used simutaneously, then the combined model can be used for finetuning using my Grantha data. I am not sure whether it will work. What is your opinion? Is this the only way like finetuning each model separately and then use this command:
"tesseract image.png -l grantha+grantha1"
The text was updated successfully, but these errors were encountered: