Image processing for OCR with leptonica (inverse c

2019-02-01 01:53发布

I am trying to process the following image with leptonica to extract text with tesseract.

Original Image: original image

Tesseract on the original image yields this:

i s l
D2J1FiiE-l191x1iitmwii9 uhiaiislz-2 Q ~37
Bottom linez
With a little time!
you can learn social media technology
using free online resources-
And if you donity
youlll be at a significant disadvantage
to
other HOn-pFOiiTS-

Not great, especially the top background. So using leptionica I use a background removal algorithm (blur, difference, threshold, invert) to get the following image: processed image

But tesseract doesn't do a good job with it:

@@r-mair lkrm@W lh@w ilr@ mJs@ iklh@ ii@c2lhm1@ll
mm Mime
VWU1 a Mitt-Jle time-
@1m ll@@Wn Om @@@lh1
using free onhne resources-
Andifyoudoni
9110 ate a $0 D
to other non-profrts
I

The main problem, it seems, is that now all of the text is outlined instead of solid. How can I adjust my algorithm or what can I add to made the text solid?

1条回答
虎瘦雄心在
2楼-- · 2019-02-01 02:39

It seems that this paper proposes a binarization method which solves your problem:

T Kasar, J Kumar and A G Ramakrishnan. Font and Background Color Independent Text Binarization. (2007)

Kasar etal method performance

查看更多
登录 后发表回答