Preface: "There is very little practice or theory in this talk."
There is a lot of discussion in certain areas of the blogsphere regarding the problems of text-image linking. Hugh has an upcoming huge manuscript digitization projects and might be able to use this theory/practice.
Goals:
- Create an SVG overlay of the manuscript page image
- Analyse the structure of the SVG document to detect lines, etc.
- Link the groups so produced to structures in a TEI transcription
- Display the results in a usable GUI
He gathered tools from the open source commnity rather than writing his own monolithic thing.
potrace - takes bitmap and converts to a vector graphic image - the image has to be bitmap (using imageMagick) The output from potrace was not what he wanted exactly so he used Inkscape from the command line to convert to absolute coordinates. Then he used XSLT to do a little cleanup (adding specifically named IDs)
He then used lxml - ElementTree and numpy - the script reads in the SVG produced by potrace and filtered through Inkscape, does some filtering,detects the lines, then serializes the results back to SVG and Javascript.
Problems:
- How much can be automated? Now not much is.
- How deeply can this be analysed?
- What is the best testing mechanism?
His process is very preliminary but really sounds fascinating to me. I have tons of documents from m grandmothers' homes that I am in fear of losing....to either the elements or the trashcan. Not everyone in the family finds them useful. But if I could digitize them, at least I could maintain the content, if not the source material. I will have to look into aspects of this process.
No comments:
Post a Comment