2010-01-17 18:10:35 +06:00
|
|
|
#! /usr/bin/python
|
|
|
|
# vim:ts=4:sw=4:softtabstop=4:smarttab:expandtab
|
|
|
|
|
|
|
|
from __future__ import with_statement
|
|
|
|
import csv
|
|
|
|
import sys
|
|
|
|
import os
|
|
|
|
import getopt
|
|
|
|
from struct import pack
|
|
|
|
from struct import unpack
|
|
|
|
|
|
|
|
|
|
|
|
class DocParser(object):
|
2010-01-19 18:11:59 +06:00
|
|
|
def __init__(self, flatxml, classlst, fileid):
|
2010-01-17 18:10:35 +06:00
|
|
|
self.id = os.path.basename(fileid).replace('.dat','')
|
2010-01-21 18:14:31 +06:00
|
|
|
self.docList = flatxml.split('\n')
|
|
|
|
self.docSize = len(self.docList)
|
2010-01-19 18:11:59 +06:00
|
|
|
self.classList = {}
|
|
|
|
tmpList = classlst.split('\n')
|
|
|
|
for pclass in tmpList:
|
|
|
|
if pclass != '':
|
|
|
|
# remove the leading period from the css name
|
|
|
|
cname = pclass[1:]
|
|
|
|
self.classList[cname] = True
|
2010-01-17 18:10:35 +06:00
|
|
|
self.ocrtext = []
|
|
|
|
self.link_id = []
|
|
|
|
self.link_title = []
|
|
|
|
self.link_page = []
|
|
|
|
self.dehyphen_rootid = []
|
|
|
|
self.paracont_stemid = []
|
|
|
|
self.parastems_stemid = []
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# return tag at line pos in document
|
2010-01-19 18:11:59 +06:00
|
|
|
def lineinDoc(self, pos) :
|
2010-01-21 18:14:31 +06:00
|
|
|
if (pos >= 0) and (pos < self.docSize) :
|
|
|
|
item = self.docList[pos]
|
2010-01-19 18:11:59 +06:00
|
|
|
if item.find('=') >= 0:
|
|
|
|
(name, argres) = item.split('=',1)
|
|
|
|
else :
|
|
|
|
name = item
|
|
|
|
argres = ''
|
|
|
|
return name, argres
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# find tag in doc if within pos to end inclusive
|
2010-01-17 18:10:35 +06:00
|
|
|
def findinDoc(self, tagpath, pos, end) :
|
|
|
|
result = None
|
|
|
|
if end == -1 :
|
2010-01-21 18:14:31 +06:00
|
|
|
end = self.docSize
|
2010-01-17 18:10:35 +06:00
|
|
|
else:
|
2010-01-21 18:14:31 +06:00
|
|
|
end = min(self.docSize, end)
|
2010-01-17 18:10:35 +06:00
|
|
|
foundat = -1
|
|
|
|
for j in xrange(pos, end):
|
2010-01-21 18:14:31 +06:00
|
|
|
item = self.docList[j]
|
2010-01-17 18:10:35 +06:00
|
|
|
if item.find('=') >= 0:
|
2010-01-21 18:14:31 +06:00
|
|
|
(name, argres) = item.split('=',1)
|
2010-01-17 18:10:35 +06:00
|
|
|
else :
|
|
|
|
name = item
|
|
|
|
argres = ''
|
|
|
|
if name.endswith(tagpath) :
|
|
|
|
result = argres
|
|
|
|
foundat = j
|
|
|
|
break
|
|
|
|
return foundat, result
|
|
|
|
|
|
|
|
|
|
|
|
# return list of start positions for the tagpath
|
|
|
|
def posinDoc(self, tagpath):
|
|
|
|
startpos = []
|
|
|
|
pos = 0
|
|
|
|
res = ""
|
|
|
|
while res != None :
|
|
|
|
(foundpos, res) = self.findinDoc(tagpath, pos, -1)
|
|
|
|
if res != None :
|
|
|
|
startpos.append(foundpos)
|
|
|
|
pos = foundpos + 1
|
|
|
|
return startpos
|
|
|
|
|
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
# build a description of the paragraph
|
2010-01-17 18:10:35 +06:00
|
|
|
def getParaDescription(self, start, end):
|
2010-01-19 18:11:59 +06:00
|
|
|
|
|
|
|
result = []
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# paragraph
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, pclass) = self.findinDoc('paragraph.class',start,end)
|
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
# class names are an issue given topaz may start them with numerals (not allowed),
|
|
|
|
# use a mix of cases (which cause some browsers problems), and actually
|
2010-01-20 18:13:31 +06:00
|
|
|
# attach numbers after "_reclustered*" to the end to deal classeses that inherit
|
|
|
|
# from a base class (but then not actually provide all of these _reclustereed
|
|
|
|
# classes in the stylesheet!
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# so we clean this up by lowercasing, prepend 'cl-', and getting any baseclass
|
2010-01-20 18:13:31 +06:00
|
|
|
# that exists in the stylesheet first, and then adding this specific class
|
|
|
|
# after
|
2010-01-21 18:14:31 +06:00
|
|
|
if pclass != None :
|
|
|
|
classres = ''
|
|
|
|
pclass = pclass.lower()
|
|
|
|
pclass = 'cl-' + pclass
|
|
|
|
p = pclass.find('_')
|
|
|
|
if p > 0 :
|
|
|
|
baseclass = pclass[0:p]
|
|
|
|
if baseclass in self.classList:
|
|
|
|
classres += baseclass + ' '
|
|
|
|
classres += pclass
|
|
|
|
pclass = classres
|
2010-01-19 18:11:59 +06:00
|
|
|
|
|
|
|
# build up a description of the paragraph in result and return it
|
|
|
|
# first check for the basic - all words paragraph
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, sfirst) = self.findinDoc('paragraph.firstWord',start,end)
|
|
|
|
(pos, slast) = self.findinDoc('paragraph.lastWord',start,end)
|
|
|
|
if (sfirst != None) and (slast != None) :
|
2010-01-19 18:11:59 +06:00
|
|
|
first = int(sfirst)
|
|
|
|
last = int(slast)
|
|
|
|
for wordnum in xrange(first, last):
|
|
|
|
result.append(('ocr', wordnum))
|
|
|
|
return pclass, result
|
|
|
|
|
|
|
|
# this type of paragrph may be made up of multiple _spans, inline
|
|
|
|
# word monograms (images) and words with semantic meaning
|
|
|
|
|
|
|
|
# need to parse this type line by line
|
|
|
|
line = start + 1
|
|
|
|
word_class = ''
|
|
|
|
|
2010-01-20 18:13:31 +06:00
|
|
|
# if end is -1 then we must search to end of document
|
|
|
|
if end == -1 :
|
2010-01-21 18:14:31 +06:00
|
|
|
end = self.docSize
|
2010-01-20 18:13:31 +06:00
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
while (line < end) :
|
|
|
|
|
|
|
|
(name, argres) = self.lineinDoc(line)
|
|
|
|
|
|
|
|
if name.endswith('_span.firstWord') :
|
|
|
|
first = int(argres)
|
|
|
|
(name, argres) = self.lineinDoc(line+1)
|
|
|
|
if not name.endswith('_span.lastWord'):
|
|
|
|
print 'Error: - incorrect _span ordering inside paragraph'
|
|
|
|
last = int(argres)
|
|
|
|
for wordnum in xrange(first, last):
|
|
|
|
result.append(('ocr', wordnum))
|
|
|
|
line += 1
|
|
|
|
|
|
|
|
elif name.endswith('word.class'):
|
|
|
|
(cname, space) = argres.split('-',1)
|
2010-01-20 18:13:31 +06:00
|
|
|
if space == '' : space = '0'
|
|
|
|
if (cname == 'spaceafter') and (int(space) > 0) :
|
2010-01-19 18:11:59 +06:00
|
|
|
word_class = 'sa'
|
|
|
|
|
|
|
|
elif name.endswith('word.img.src'):
|
|
|
|
result.append(('img' + word_class, int(argres)))
|
|
|
|
word_class = ''
|
|
|
|
|
|
|
|
elif name.endswith('word_semantic.firstWord'):
|
|
|
|
first = int(argres)
|
|
|
|
(name, argres) = self.lineinDoc(line+1)
|
|
|
|
if not name.endswith('word_semantic.lastWord'):
|
|
|
|
print 'Error: - incorrect word_semantic ordering inside paragraph'
|
|
|
|
last = int(argres)
|
|
|
|
for wordnum in xrange(first, last):
|
|
|
|
result.append(('ocr', wordnum))
|
|
|
|
line += 1
|
|
|
|
|
|
|
|
line += 1
|
|
|
|
|
|
|
|
return pclass, result
|
|
|
|
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
def buildParagraph(self, pclass, pdesc, type, regtype) :
|
2010-01-17 18:10:35 +06:00
|
|
|
parares = ''
|
|
|
|
sep =''
|
2010-01-19 18:11:59 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
classres = ''
|
|
|
|
if pclass :
|
|
|
|
classres = ' class="' + pclass + '"'
|
2010-01-19 18:11:59 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
br_lb = (regtype == 'fixed') or (regtype == 'chapterheading')
|
2010-01-19 18:11:59 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
handle_links = len(self.link_id) > 0
|
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
if (type == 'full') or (type == 'begin') :
|
2010-01-21 18:14:31 +06:00
|
|
|
parares += '<p' + classres + '>'
|
2010-01-19 18:11:59 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
if (type == 'end'):
|
|
|
|
parares += ' '
|
2010-01-19 18:11:59 +06:00
|
|
|
|
|
|
|
cnt = len(pdesc)
|
|
|
|
|
|
|
|
for j in xrange( 0, cnt) :
|
|
|
|
|
|
|
|
(wtype, num) = pdesc[j]
|
|
|
|
|
|
|
|
if wtype == 'ocr' :
|
|
|
|
word = self.ocrtext[num]
|
|
|
|
sep = ' '
|
|
|
|
|
|
|
|
if handle_links:
|
|
|
|
link = self.link_id[num]
|
|
|
|
if (link > 0):
|
|
|
|
title = self.link_title[link-1]
|
2010-01-20 18:13:31 +06:00
|
|
|
if (title == "") or (parares.rfind(title) < 0):
|
|
|
|
title='_link_'
|
2010-01-19 18:11:59 +06:00
|
|
|
ptarget = self.link_page[link-1] - 1
|
|
|
|
linkhtml = '<a href="#page%04d">' % ptarget
|
|
|
|
linkhtml += title + '</a>'
|
|
|
|
pos = parares.rfind(title)
|
|
|
|
if pos >= 0:
|
|
|
|
parares = parares[0:pos] + linkhtml + parares[pos+len(title):]
|
|
|
|
else :
|
|
|
|
parares += linkhtml
|
|
|
|
if word == '_link_' : word = ''
|
|
|
|
elif (link < 0) :
|
|
|
|
if word == '_link_' : word = ''
|
|
|
|
|
|
|
|
if word == '_lb_':
|
2010-01-21 18:14:31 +06:00
|
|
|
if ((num-1) in self.dehyphen_rootid ) or handle_links:
|
2010-01-19 18:11:59 +06:00
|
|
|
word = ''
|
|
|
|
sep = ''
|
|
|
|
elif br_lb :
|
|
|
|
word = '<br />\n'
|
|
|
|
sep = ''
|
2010-01-17 18:10:35 +06:00
|
|
|
else :
|
2010-01-19 18:11:59 +06:00
|
|
|
word = '\n'
|
|
|
|
sep = ''
|
|
|
|
|
|
|
|
if num in self.dehyphen_rootid :
|
|
|
|
word = word[0:-1]
|
2010-01-17 18:10:35 +06:00
|
|
|
sep = ''
|
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
parares += word + sep
|
|
|
|
|
|
|
|
elif wtype == 'img' :
|
2010-01-17 18:10:35 +06:00
|
|
|
sep = ''
|
2010-01-19 18:11:59 +06:00
|
|
|
parares += '<img src="img/img%04d.jpg" alt="" />' % num
|
|
|
|
parares += sep
|
2010-01-17 18:10:35 +06:00
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
elif wtype == 'imgsa' :
|
|
|
|
sep = ' '
|
|
|
|
parares += '<img src="img/img%04d.jpg" alt="" />' % num
|
|
|
|
parares += sep
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
if len(sep) > 0 : parares = parares[0:-1]
|
|
|
|
if (type == 'full') or (type == 'end') :
|
|
|
|
parares += '</p>'
|
|
|
|
return parares
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# walk the document tree collecting the information needed
|
|
|
|
# to build an html page using the ocrText
|
|
|
|
|
|
|
|
def process(self):
|
|
|
|
|
|
|
|
htmlpage = ''
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# get the ocr text
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, argres) = self.findinDoc('info.word.ocrText',0,-1)
|
|
|
|
if argres : self.ocrtext = argres.split('|')
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# get information to dehyphenate the text
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, argres) = self.findinDoc('info.dehyphen.rootID',0,-1)
|
|
|
|
if argres:
|
|
|
|
argList = argres.split('|')
|
|
|
|
self.dehyphen_rootid = [ int(strval) for strval in argList]
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# determine if first paragraph is continued from previous page
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, self.parastems_stemid) = self.findinDoc('info.paraStems.stemID',0,-1)
|
2010-01-21 18:14:31 +06:00
|
|
|
first_para_continued = (self.parastems_stemid != None)
|
|
|
|
|
|
|
|
# determine if last paragraph is continued onto the next page
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, self.paracont_stemid) = self.findinDoc('info.paraCont.stemID',0,-1)
|
2010-01-21 18:14:31 +06:00
|
|
|
last_para_continued = (self.paracont_stemid != None)
|
2010-01-17 18:10:35 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# collect link ids
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, argres) = self.findinDoc('info.word.link_id',0,-1)
|
|
|
|
if argres:
|
|
|
|
argList = argres.split('|')
|
|
|
|
self.link_id = [ int(strval) for strval in argList]
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# collect link destination page numbers
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, argres) = self.findinDoc('info.links.page',0,-1)
|
|
|
|
if argres :
|
|
|
|
argList = argres.split('|')
|
|
|
|
self.link_page = [ int(strval) for strval in argList]
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# collect link titles
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, argres) = self.findinDoc('info.links.title',0,-1)
|
|
|
|
if argres :
|
|
|
|
self.link_title = argres.split('|')
|
|
|
|
else:
|
|
|
|
self.link_title.append('')
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
|
|
|
# get page type
|
2010-01-17 18:10:35 +06:00
|
|
|
(pos, pagetype) = self.findinDoc('page.type',0,-1)
|
|
|
|
|
|
|
|
|
|
|
|
# generate a list of each region starting point
|
|
|
|
# each region has one paragraph,, or one image, or one chapterheading
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
regionList= self.posinDoc('region')
|
|
|
|
regcnt = len(regionList)
|
|
|
|
regionList.append(-1)
|
|
|
|
|
|
|
|
anchorSet = False
|
|
|
|
breakSet = False
|
|
|
|
|
|
|
|
# process each region tag and convert what you can to html
|
|
|
|
|
|
|
|
for j in xrange(regcnt):
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
start = regionList[j]
|
|
|
|
end = regionList[j+1]
|
|
|
|
|
|
|
|
(pos, regtype) = self.findinDoc('region.type',start,end)
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
# set anchor for link target on this page
|
|
|
|
if not anchorSet and not first_para_continued:
|
|
|
|
htmlpage += '<div id="' + self.id + '" class="page_' + pagetype + '"> </div>\n'
|
|
|
|
anchorSet = True
|
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
if regtype == 'graphic' :
|
|
|
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
|
|
|
if simgsrc:
|
|
|
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
elif regtype == 'chapterheading' :
|
2010-01-19 18:11:59 +06:00
|
|
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
2010-01-17 18:10:35 +06:00
|
|
|
if not breakSet:
|
|
|
|
htmlpage += '<div style="page-break-after: always;"> </div>\n'
|
|
|
|
breakSet = True
|
|
|
|
tag = 'h1'
|
2010-01-21 18:14:31 +06:00
|
|
|
if pclass and (len(pclass) >= 7):
|
|
|
|
if pclass[3:7] == 'ch1-' : tag = 'h1'
|
|
|
|
if pclass[3:7] == 'ch2-' : tag = 'h2'
|
|
|
|
if pclass[3:7] == 'ch3-' : tag = 'h3'
|
|
|
|
htmlpage += '<' + tag + ' class="' + pclass + '">'
|
|
|
|
else:
|
|
|
|
htmlpage += '<' + tag + '>'
|
2010-01-19 18:11:59 +06:00
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
|
2010-01-17 18:10:35 +06:00
|
|
|
htmlpage += '</' + tag + '>'
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-20 18:13:31 +06:00
|
|
|
elif (regtype == 'text') or (regtype == 'fixed') or (regtype == 'insert') or (regtype == 'listitem') :
|
2010-01-17 18:10:35 +06:00
|
|
|
ptype = 'full'
|
|
|
|
# check to see if this is a continution from the previous page
|
2010-01-21 18:14:31 +06:00
|
|
|
if first_para_continued :
|
2010-01-17 18:10:35 +06:00
|
|
|
ptype = 'end'
|
2010-01-21 18:14:31 +06:00
|
|
|
first_para_continued = False
|
2010-01-19 18:11:59 +06:00
|
|
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
2010-01-21 18:14:31 +06:00
|
|
|
if pclass and (len(pclass) >= 6) and (ptype == 'full'):
|
2010-01-17 18:10:35 +06:00
|
|
|
tag = 'p'
|
|
|
|
if pclass[3:6] == 'h1-' : tag = 'h4'
|
|
|
|
if pclass[3:6] == 'h2-' : tag = 'h5'
|
|
|
|
if pclass[3:6] == 'h3-' : tag = 'h6'
|
|
|
|
htmlpage += '<' + tag + ' class="' + pclass + '">'
|
2010-01-19 18:11:59 +06:00
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
|
2010-01-17 18:10:35 +06:00
|
|
|
htmlpage += '</' + tag + '>'
|
|
|
|
else :
|
2010-01-19 18:11:59 +06:00
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
2010-01-17 18:10:35 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
elif (regtype == 'tocentry') :
|
|
|
|
ptype = 'full'
|
2010-01-21 18:14:31 +06:00
|
|
|
if first_para_continued :
|
2010-01-17 18:10:35 +06:00
|
|
|
ptype = 'end'
|
2010-01-21 18:14:31 +06:00
|
|
|
first_para_continued = False
|
2010-01-19 18:11:59 +06:00
|
|
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-20 18:13:31 +06:00
|
|
|
elif (regtype == 'synth_fcvr.center') or (regtype == 'synth_text.center'):
|
2010-01-19 18:11:59 +06:00
|
|
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
|
|
|
if simgsrc:
|
|
|
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
2010-01-17 18:10:35 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
2010-01-17 18:10:35 +06:00
|
|
|
else :
|
2010-01-19 18:11:59 +06:00
|
|
|
print 'Warning: Unknown region type', regtype
|
2010-01-20 18:13:31 +06:00
|
|
|
(pos, temp) = self.findinDoc('paragraph',start,end)
|
|
|
|
if temp:
|
|
|
|
print 'Treating this like a "text" region'
|
|
|
|
regtype = 'fixed'
|
|
|
|
ptype = 'full'
|
|
|
|
# check to see if this is a continution from the previous page
|
2010-01-21 18:14:31 +06:00
|
|
|
if first_para_continued :
|
2010-01-20 18:13:31 +06:00
|
|
|
ptype = 'end'
|
2010-01-21 18:14:31 +06:00
|
|
|
first_para_continued = False
|
2010-01-20 18:13:31 +06:00
|
|
|
(pclass, pdesc) = self.getParaDescription(start,end)
|
2010-01-21 18:14:31 +06:00
|
|
|
if pclass and (ptype == 'full') and (len(pclass) >= 6):
|
2010-01-20 18:13:31 +06:00
|
|
|
tag = 'p'
|
|
|
|
if pclass[3:6] == 'h1-' : tag = 'h4'
|
|
|
|
if pclass[3:6] == 'h2-' : tag = 'h5'
|
|
|
|
if pclass[3:6] == 'h3-' : tag = 'h6'
|
|
|
|
htmlpage += '<' + tag + ' class="' + pclass + '">'
|
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
|
|
|
|
htmlpage += '</' + tag + '>'
|
|
|
|
else :
|
|
|
|
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
|
|
|
|
else :
|
2010-01-21 18:14:31 +06:00
|
|
|
print 'Treating this like a "graphic" region'
|
2010-01-20 18:13:31 +06:00
|
|
|
(pos, simgsrc) = self.findinDoc('img.src',start,end)
|
|
|
|
if simgsrc:
|
|
|
|
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
|
2010-01-17 18:10:35 +06:00
|
|
|
|
2010-01-21 18:14:31 +06:00
|
|
|
|
|
|
|
if last_para_continued :
|
2010-01-17 18:10:35 +06:00
|
|
|
if htmlpage[-4:] == '</p>':
|
2010-01-21 18:14:31 +06:00
|
|
|
htmlpage = htmlpage[0:-4]
|
|
|
|
last_para_continued = False
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
return htmlpage
|
|
|
|
|
|
|
|
|
|
|
|
|
2010-01-19 18:11:59 +06:00
|
|
|
def convert2HTML(flatxml, classlst, fileid):
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
# create a document parser
|
2010-01-19 18:11:59 +06:00
|
|
|
dp = DocParser(flatxml, classlst, fileid)
|
2010-01-17 18:10:35 +06:00
|
|
|
|
|
|
|
htmlpage = dp.process()
|
|
|
|
|
|
|
|
return htmlpage
|