mirror of
https://github.com/qurator-spk/eynollah.git
synced 2025-06-30 22:49:52 +02:00
Update README.md
This commit is contained in:
parent
702656cf77
commit
757987314d
1 changed files with 14 additions and 6 deletions
20
README.md
20
README.md
|
@ -17,13 +17,14 @@
|
||||||
* Detection of reading order
|
* Detection of reading order
|
||||||
* Output in [PAGE-XML](https://github.com/PRImA-Research-Lab/PAGE-XML)
|
* Output in [PAGE-XML](https://github.com/PRImA-Research-Lab/PAGE-XML)
|
||||||
* [OCR-D](https://github.com/qurator-spk/eynollah#use-as-ocr-d-processor) interface
|
* [OCR-D](https://github.com/qurator-spk/eynollah#use-as-ocr-d-processor) interface
|
||||||
|
* [Examples](https://github.com/qurator-spk/eynollah/wiki#examples)
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
Python versions `3.8-3.11` with Tensorflow versions >=`2.12` on Linux are currently supported. While we can not provide support for Windows or MacOS, Windows users may be able to install and run the tool through Linux in [WSL](https://learn.microsoft.com/en-us/windows/wsl/).
|
Python versions `3.8-3.11` with Tensorflow versions >=`2.12` on Linux are currently supported.
|
||||||
|
|
||||||
For (limited) GPU support the CUDA toolkit needs to be installed.
|
For (limited) GPU support the CUDA toolkit needs to be installed.
|
||||||
|
|
||||||
You can either install from PyPI via
|
You can either install from PyPI
|
||||||
|
|
||||||
```
|
```
|
||||||
pip install eynollah
|
pip install eynollah
|
||||||
|
@ -39,9 +40,14 @@ cd eynollah; pip install -e .
|
||||||
Alternatively, run `make install` or `make install-dev` for editable installation.
|
Alternatively, run `make install` or `make install-dev` for editable installation.
|
||||||
|
|
||||||
## Models
|
## Models
|
||||||
Pre-trained models can be downloaded from [qurator-data.de](https://qurator-data.de/eynollah/). In case you want to train your own model with Eynollah, have a look at [`train`](https://github.com/qurator-spk/eynollah/tree/main/eynollah/eynollah/train).
|
Pre-trained models can be downloaded either from [qurator-data.de](https://qurator-data.de/eynollah/) or [huggingface](https://huggingface.co/SBB).
|
||||||
|
|
||||||
## Usage
|
## Train
|
||||||
|
🚧 **Work in progress**
|
||||||
|
|
||||||
|
In case you want to train your own model, have a look at [`train`](https://github.com/qurator-spk/eynollah/tree/main/eynollah/eynollah/train).
|
||||||
|
|
||||||
|
## Use
|
||||||
The command-line interface can be called like this:
|
The command-line interface can be called like this:
|
||||||
|
|
||||||
```sh
|
```sh
|
||||||
|
@ -76,7 +82,6 @@ If no option is set, the tool will perform layout detection of main regions (bac
|
||||||
The tool produces better quality output when RGB images are used as input than greyscale or binarized images.
|
The tool produces better quality output when RGB images are used as input than greyscale or binarized images.
|
||||||
|
|
||||||
#### Use as OCR-D processor
|
#### Use as OCR-D processor
|
||||||
|
|
||||||
Eynollah ships with a CLI interface to be used as [OCR-D](https://ocr-d.de) processor.
|
Eynollah ships with a CLI interface to be used as [OCR-D](https://ocr-d.de) processor.
|
||||||
|
|
||||||
In this case, the source image file group with (preferably) RGB images should be used as input like this:
|
In this case, the source image file group with (preferably) RGB images should be used as input like this:
|
||||||
|
@ -93,8 +98,11 @@ ocrd-eynollah-segment -I OCR-D-IMG-BIN -O SEG-LINE -P models
|
||||||
|
|
||||||
uses the original (RGB) image despite any binarization that may have occured in previous OCR-D processing steps
|
uses the original (RGB) image despite any binarization that may have occured in previous OCR-D processing steps
|
||||||
|
|
||||||
|
#### Additional documentation
|
||||||
|
Please check the [wiki](https://github.com/qurator-spk/eynollah/wiki).
|
||||||
|
|
||||||
## How to cite
|
## How to cite
|
||||||
If you find this tool useful in your work, please consider citing our paper:
|
If you find this useful in your work, please consider citing our paper:
|
||||||
|
|
||||||
```bibtex
|
```bibtex
|
||||||
@inproceedings{rezanezhad2023eynollah,
|
@inproceedings{rezanezhad2023eynollah,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue