You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
Go to file
Konstantin Baierer e0d38517d3
Merge pull request #130 from qurator-spk/v3-api
port processor to core v3
5 days ago
.github/workflows CI: add binarization models to cache 1 week ago
src/eynollah In light mode: To determine whether a main region is a header, I adjusted the ratio to achieve better results. 5 days ago
tests smoke-test: also test dir-in mode and overwrite 1 week ago
.dockerignore add Continuous Deployment via Dockerhub and GHCR 1 week ago
.gitignore 📦 v0.0.3 4 years ago
CHANGELOG.md 📦 v0.3.1 8 months ago
Dockerfile dockerfile: add smoke test 1 week ago
LICENSE extend setup.py, add Makefile, gitignore, requirements.txt 4 years ago
Makefile smoke-test: also test dir-in mode and overwrite 1 week ago
README.md another fix to avoid frequent warnings 1 week ago
ocrd-tool.json switch from qurator namespace to src-layout 7 months ago
pyproject.toml improve+extend makefile 1 week ago
requirements-test.txt 🎨 reformat cli.py with black 4 years ago
requirements.txt Update requirements.txt 1 week ago

README.md

Eynollah

Document Layout Analysis with Deep Learning and Heuristics

PyPI Version GH Actions Test License: ASL DOI

Features

  • Support for up to 10 segmentation classes:
  • Support for various image optimization operations:
    • cropping (border detection), binarization, deskewing, dewarping, scaling, enhancing, resizing
  • Text line segmentation to bounding boxes or polygons (contours) including for curved lines and vertical text
  • Detection of reading order (left-to-right or right-to-left)
  • Output in PAGE-XML
  • OCR-D interface

⚠️ Development is currently focused on achieving the best possible quality of results for a wide variety of historical documents and therefore processing can be very slow. We aim to improve this, but contributions are welcome.

Installation

Python 3.8-3.11 with Tensorflow <2.13 on Linux are currently supported.

For (limited) GPU support the CUDA toolkit needs to be installed.

You can either install from PyPI

pip install eynollah

or clone the repository, enter it and install (editable) with

git clone git@github.com:qurator-spk/eynollah.git
cd eynollah; pip install -e .

Alternatively, you can run make install or make install-dev for editable installation.

Models

Pre-trained models can be downloaded from qurator-data.de or huggingface.

Train

🚧 Work in progress

In case you want to train your own model, have a look at sbb_pixelwise_segmentation.

Usage

The command-line interface can be called like this:

eynollah \
  -i <single image file> | -di <directory containing image files> \
  -o <output directory> \
  -m <directory containing model files> \
     [OPTIONS]

The following options can be used to further configure the processing:

option description
-fl full layout analysis including all steps and segmentation classes
-light lighter and faster but simpler method for main region detection and deskewing
-tab apply table detection
-ae apply enhancement (the resulting image is saved to the output directory)
-as apply scaling
-cl apply contour detection for curved text lines instead of bounding boxes
-ib apply binarization (the resulting image is saved to the output directory)
-ep enable plotting (MUST always be used with -sl, -sd, -sa, -si or -ae)
-eoi extract only images to output directory (other processing will not be done)
-ho ignore headers for reading order dectection
-si <directory> save image regions detected to this directory
-sd <directory> save deskewed image to this directory
-sl <directory> save layout prediction as plot to this directory
-sp <directory> save cropped page image to this directory
-sa <directory> save all (plot, enhanced/binary image, layout) to this directory

If no option is set, the tool performs layout detection of main regions (background, text, images, separators and marginals). The best output quality is produced when RGB images are used as input rather than greyscale or binarized images.

Use as OCR-D processor

Eynollah ships with a CLI interface to be used as OCR-D processor.

In this case, the source image file group with (preferably) RGB images should be used as input like this:

ocrd-eynollah-segment -I OCR-D-IMG -O OCR-D-SEG -P models 2022-04-05

If the input file group is PAGE-XML (from a previous OCR-D workflow step), Eynollah behaves as follows:

  • existing regions are kept and ignored (i.e. in effect they might overlap segments from Eynollah results)

  • existing annotation (and respective AlternativeImages) are partially ignored:

    • previous page frame detection (cropped images)
    • previous derotation (deskewed images)
    • previous thresholding (binarized images)
  • if the page-level image nevertheless deviates from the original (@imageFilename) (because some other preprocessing step was in effect like denoised), then the output PAGE-XML will be based on that as new top-level (@imageFilename)

    ocrd-eynollah-segment -I OCR-D-XYZ -O OCR-D-SEG -P models 2022-04-05

Still, in general, it makes more sense to add other workflow steps after Eynollah.

Additional documentation

Please check the wiki.

How to cite

If you find this tool useful in your work, please consider citing our paper:

@inproceedings{hip23rezanezhad,
  title     = {Document Layout Analysis with Deep Learning and Heuristics},
  author    = {Rezanezhad, Vahid and Baierer, Konstantin and Gerber, Mike and Labusch, Kai and Neudecker, Clemens},
  booktitle = {Proceedings of the 7th International Workshop on Historical Document Imaging and Processing {HIP} 2023,
               San José, CA, USA, August 25-26, 2023},
  publisher = {Association for Computing Machinery},
  address   = {New York, NY, USA},
  year      = {2023},
  pages     = {73--78},
  url       = {https://doi.org/10.1145/3604951.3605513}
}