9f875fb819
This fixes a name conflict for Lao.traineddata and lao.traineddata which could not be distinguished on case insensitive filesystems (for example macOS, Windows). It makes it also easier for users to see which data is for scripts. Choosing a script works now like this: tesseract -l script/Latin ... Signed-off-by: Stefan Weil <sw@weilnetz.de> |
||
---|---|---|
script | ||
afr.traineddata | ||
amh.traineddata | ||
ara.traineddata | ||
asm.traineddata | ||
aze.traineddata | ||
aze_cyrl.traineddata | ||
bel.traineddata | ||
ben.traineddata | ||
bod.traineddata | ||
bos.traineddata | ||
bre.traineddata | ||
bul.traineddata | ||
cat.traineddata | ||
ceb.traineddata | ||
ces.traineddata | ||
chi_sim.traineddata | ||
chi_sim_vert.traineddata | ||
chi_tra.traineddata | ||
chi_tra_vert.traineddata | ||
chr.traineddata | ||
COPYING | ||
cos.traineddata | ||
cym.traineddata | ||
dan.traineddata | ||
deu.traineddata | ||
div.traineddata | ||
dzo.traineddata | ||
ell.traineddata | ||
eng.traineddata | ||
enm.traineddata | ||
epo.traineddata | ||
est.traineddata | ||
eus.traineddata | ||
fao.traineddata | ||
fas.traineddata | ||
fil.traineddata | ||
fin.traineddata | ||
fra.traineddata | ||
frk.traineddata | ||
frm.traineddata | ||
fry.traineddata | ||
gla.traineddata | ||
gle.traineddata | ||
glg.traineddata | ||
grc.traineddata | ||
guj.traineddata | ||
hat.traineddata | ||
heb.traineddata | ||
hin.traineddata | ||
hrv.traineddata | ||
hun.traineddata | ||
hye.traineddata | ||
iku.traineddata | ||
ind.traineddata | ||
isl.traineddata | ||
ita.traineddata | ||
ita_old.traineddata | ||
jav.traineddata | ||
jpn.traineddata | ||
jpn_vert.traineddata | ||
kan.traineddata | ||
kat.traineddata | ||
kat_old.traineddata | ||
kaz.traineddata | ||
khm.traineddata | ||
kir.traineddata | ||
kor.traineddata | ||
kor_vert.traineddata | ||
kur_ara.traineddata | ||
lao.traineddata | ||
lat.traineddata | ||
lav.traineddata | ||
lit.traineddata | ||
ltz.traineddata | ||
mal.traineddata | ||
mar.traineddata | ||
mkd.traineddata | ||
mlt.traineddata | ||
mon.traineddata | ||
mri.traineddata | ||
msa.traineddata | ||
mya.traineddata | ||
nep.traineddata | ||
nld.traineddata | ||
nor.traineddata | ||
oci.traineddata | ||
ori.traineddata | ||
osd.traineddata | ||
pan.traineddata | ||
pol.traineddata | ||
por.traineddata | ||
pus.traineddata | ||
que.traineddata | ||
README.md | ||
ron.traineddata | ||
rus.traineddata | ||
san.traineddata | ||
sin.traineddata | ||
slk.traineddata | ||
slv.traineddata | ||
snd.traineddata | ||
spa.traineddata | ||
spa_old.traineddata | ||
sqi.traineddata | ||
srp.traineddata | ||
srp_latn.traineddata | ||
sun.traineddata | ||
swa.traineddata | ||
swe.traineddata | ||
syr.traineddata | ||
tam.traineddata | ||
tat.traineddata | ||
tel.traineddata | ||
tgk.traineddata | ||
tha.traineddata | ||
tir.traineddata | ||
ton.traineddata | ||
tur.traineddata | ||
uig.traineddata | ||
ukr.traineddata | ||
urd.traineddata | ||
uzb.traineddata | ||
uzb_cyrl.traineddata | ||
vie.traineddata | ||
yid.traineddata | ||
yor.traineddata |
tessdata_fast – Fast integer versions of trained models
This repository contains fast integer versions of trained models for the Tesseract Open Source OCR Engine.
- Most users will want to use these traineddata files to do OCR and these will be shipped as part of Linux distributions.
- Fine tuning/incremental training will NOT be possible from these
fast
models, as they are 8-bit integer. - It will be possible to convert a tuned
best
to integer to make it faster, but some of the speed infast
will be from the smaller model. - When using the models in this repository, only the new LSTM-based OCR engine is supported. The legacy
tesseract
engine is not supported with these files, so Tesseract's oem modes '0' and '2' won't work with them.
Two types of models
The repository contains two types of models,
- those for a single language and
- those for a single script supporting one or more languages.
Most of the script models include English training data as well as the script, but not Cyrillic, as that would have a major ambiguity problem.
On Linux, the language based traineddata packages are named tesseract-ocr-LANG
where LANG is the three letter language code eg. tesseract-ocr-eng (English language), tesseract-ocr-hin (Hindi language), etc.
On Linux, the script based traineddata packages are named tesseract-ocr-script-SCRIPT
where SCRIPT is the four letter script code eg. tesseract-ocr-script-latn (Latin Script), tesseract-ocr-script-deva (Devanagari Script), etc.
Data files for a particular script
Initial capitals in the filename indicate the one model for all languages in that script.
- Latin is all latin-based languages, except vie.
- Vietnamese is for latin-based Vietnamese language.
- Fraktur is basically a combination of all the latin-based languages that have an 'old' variant.
- Devanagari is for hin+san+mar+nep+eng.
LSTM training details for different languages and scripts
For Latin-based languages, the existing model data provided has been trained on about 400000 textlines spanning about 4500 fonts. For other scripts, not so many fonts are available, but they have still been trained on a similar number of textlines. eg.
- Latin ~4500 fonts
- Devanagari ~50 fonts
- Kannada 15.
With a theory that poor accuracy on test data and over-fitting on training data was caused by the lack of fonts, the training data has been mixed with English, so that some of the font diversity might generalize to the other script. The overall effect was slightly positive, hence the script models include English language also.
Example - jpn and Japanese
'jpn' contains whatever appears on the www that is labelled as the language, trained only with fonts that can render Japanese.
Japanese contains all the languages that use that script (in this case just the one) PLUS English.The resulting model is trained with a mix of both training sets, with the expectation that some of the generalization to 4500 English training fonts will also apply to the other script that has a lot less.
'jpn_vert' is trained on text rendered vertically (but the image is rotated so the long edge is still horizontal).
'jpn' loads 'jpn_vert' as a secondary language so it can try it in case the text is rendered vertically. This seems to work most of the time as a reasonable solution.
See the Tesseract wiki for additional information.
All data in the repository are licensed under the Apache-2.0 License, see file COPYING.