DeDRM_tools/Topaz_Tools/lib/flatxml2html.py

600 lines
22 KiB
Python
Raw Normal View History

2010-01-17 18:10:35 +06:00
#! /usr/bin/python
# vim:ts=4:sw=4:softtabstop=4:smarttab:expandtab
2010-01-24 18:19:20 +06:00
# For use with Topaz Scripts Version 1.8
2010-01-17 18:10:35 +06:00
from __future__ import with_statement
import csv
import sys
import os
2010-01-24 18:19:20 +06:00
import math
2010-01-17 18:10:35 +06:00
import getopt
from struct import pack
from struct import unpack
class DocParser(object):
2010-01-24 18:19:20 +06:00
def __init__(self, flatxml, classlst, fileid, bookDir):
2010-01-17 18:10:35 +06:00
self.id = os.path.basename(fileid).replace('.dat','')
2010-01-24 18:19:20 +06:00
self.svgcount = 0
2010-01-21 18:14:31 +06:00
self.docList = flatxml.split('\n')
self.docSize = len(self.docList)
2010-01-19 18:11:59 +06:00
self.classList = {}
2010-01-24 18:19:20 +06:00
self.bookDir = bookDir
self.glyphPaths = { }
self.numPaths = 0
2010-01-19 18:11:59 +06:00
tmpList = classlst.split('\n')
for pclass in tmpList:
if pclass != '':
# remove the leading period from the css name
cname = pclass[1:]
self.classList[cname] = True
2010-01-17 18:10:35 +06:00
self.ocrtext = []
self.link_id = []
self.link_title = []
self.link_page = []
self.dehyphen_rootid = []
self.paracont_stemid = []
self.parastems_stemid = []
2010-01-24 18:19:20 +06:00
def getGlyph(self, gid):
result = ''
id='gl%d' % gid
return self.glyphPaths[id]
def glyphs_to_image(self, glyphList):
def extract(path, key):
b = path.find(key) + len(key)
e = path.find(' ',b)
return int(path[b:e])
def extractID(path, key):
b = path.find(key) + len(key)
e = path.find('"',b)
return path[b:e]
svgDir = os.path.join(self.bookDir,'svg')
glyfile = os.path.join(svgDir,'glyphs.svg')
imgDir = os.path.join(self.bookDir,'img')
imgname = self.id + '_%04d.svg' % self.svgcount
imgfile = os.path.join(imgDir,imgname)
# build hash table of glyph paths keyed by glyph id
if self.numPaths == 0:
gfile = open(glyfile, 'r')
while True:
path = gfile.readline()
if (path == ''): break
glyphid = extractID(path,'id="')
self.glyphPaths[glyphid] = path
self.numPaths += 1
gfile.close()
# get glyph information
gxList = self.getData('info.glyph.x',0,-1)
gyList = self.getData('info.glyph.y',0,-1)
gidList = self.getData('info.glyph.glyphID',0,-1)
gids = []
maxws = []
maxhs = []
xs = []
ys = []
gdefs = []
# get path defintions, positions, dimensions for ecah glyph
# that makes up the image, and find min x and min y to reposition origin
minx = -1
miny = -1
for j in glyphList:
gid = gidList[j]
gids.append(gid)
xs.append(gxList[j])
if minx == -1: minx = gxList[j]
else : minx = min(minx, gxList[j])
ys.append(gyList[j])
if miny == -1: miny = gyList[j]
else : miny = min(miny, gyList[j])
path = self.getGlyph(gid)
gdefs.append(path)
maxws.append(extract(path,'width='))
maxhs.append(extract(path,'height='))
# change the origin to minx, miny and calc max height and width
maxw = maxws[0] + xs[0] - minx
maxh = maxhs[0] + ys[0] - miny
for j in xrange(0, len(xs)):
xs[j] = xs[j] - minx
ys[j] = ys[j] - miny
maxw = max( maxw, (maxws[j] + xs[j]) )
maxh = max( maxh, (maxhs[j] + ys[j]) )
# open the image file for output
ifile = open(imgfile,'w')
ifile.write('<?xml version="1.0" standalone="no"?>\n')
ifile.write('<!DOCTYPE svg PUBLIC "-//W3C/DTD SVG 1.1//EN" "http://www.w3.org/Graphics/SVG/1.1/DTD/svg11.dtd">\n')
ifile.write('<svg width="%dpx" height="%dpx" viewBox="0 0 %d %d" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" version="1.1">\n' % (math.floor(maxw/10), math.floor(maxh/10), maxw, maxh))
ifile.write('<defs>\n')
for j in xrange(0,len(gdefs)):
ifile.write(gdefs[j])
ifile.write('</defs>\n')
for j in xrange(0,len(gids)):
ifile.write('<use xlink:href="#gl%d" x="%d" y="%d" />\n' % (gids[j], xs[j], ys[j]))
ifile.write('</svg>')
ifile.close()
return 0
2010-01-21 18:14:31 +06:00
# return tag at line pos in document
2010-01-19 18:11:59 +06:00
def lineinDoc(self, pos) :
2010-01-21 18:14:31 +06:00
if (pos >= 0) and (pos < self.docSize) :
item = self.docList[pos]
2010-01-19 18:11:59 +06:00
if item.find('=') >= 0:
(name, argres) = item.split('=',1)
else :
name = item
argres = ''
return name, argres
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
# find tag in doc if within pos to end inclusive
2010-01-17 18:10:35 +06:00
def findinDoc(self, tagpath, pos, end) :
result = None
if end == -1 :
2010-01-21 18:14:31 +06:00
end = self.docSize
2010-01-17 18:10:35 +06:00
else:
2010-01-21 18:14:31 +06:00
end = min(self.docSize, end)
2010-01-17 18:10:35 +06:00
foundat = -1
for j in xrange(pos, end):
2010-01-21 18:14:31 +06:00
item = self.docList[j]
2010-01-17 18:10:35 +06:00
if item.find('=') >= 0:
2010-01-21 18:14:31 +06:00
(name, argres) = item.split('=',1)
2010-01-17 18:10:35 +06:00
else :
name = item
argres = ''
if name.endswith(tagpath) :
result = argres
foundat = j
break
return foundat, result
# return list of start positions for the tagpath
def posinDoc(self, tagpath):
startpos = []
pos = 0
res = ""
while res != None :
(foundpos, res) = self.findinDoc(tagpath, pos, -1)
if res != None :
startpos.append(foundpos)
pos = foundpos + 1
return startpos
2010-01-24 18:19:20 +06:00
# returns a vector of integers for the tagpath
def getData(self, tagpath, pos, end):
argres=[]
(foundat, argt) = self.findinDoc(tagpath, pos, end)
if (argt != None) and (len(argt) > 0) :
argList = argt.split('|')
argres = [ int(strval) for strval in argList]
return argres
2010-01-19 18:11:59 +06:00
# build a description of the paragraph
2010-01-17 18:10:35 +06:00
def getParaDescription(self, start, end):
2010-01-19 18:11:59 +06:00
result = []
2010-01-21 18:14:31 +06:00
# paragraph
2010-01-17 18:10:35 +06:00
(pos, pclass) = self.findinDoc('paragraph.class',start,end)
2010-01-19 18:11:59 +06:00
# class names are an issue given topaz may start them with numerals (not allowed),
# use a mix of cases (which cause some browsers problems), and actually
2010-01-20 18:13:31 +06:00
# attach numbers after "_reclustered*" to the end to deal classeses that inherit
# from a base class (but then not actually provide all of these _reclustereed
# classes in the stylesheet!
2010-01-21 18:14:31 +06:00
# so we clean this up by lowercasing, prepend 'cl-', and getting any baseclass
2010-01-20 18:13:31 +06:00
# that exists in the stylesheet first, and then adding this specific class
# after
2010-01-21 18:14:31 +06:00
if pclass != None :
classres = ''
pclass = pclass.lower()
pclass = 'cl-' + pclass
p = pclass.find('_')
if p > 0 :
baseclass = pclass[0:p]
if baseclass in self.classList:
classres += baseclass + ' '
classres += pclass
pclass = classres
2010-01-19 18:11:59 +06:00
# build up a description of the paragraph in result and return it
# first check for the basic - all words paragraph
2010-01-17 18:10:35 +06:00
(pos, sfirst) = self.findinDoc('paragraph.firstWord',start,end)
(pos, slast) = self.findinDoc('paragraph.lastWord',start,end)
if (sfirst != None) and (slast != None) :
2010-01-19 18:11:59 +06:00
first = int(sfirst)
last = int(slast)
for wordnum in xrange(first, last):
result.append(('ocr', wordnum))
return pclass, result
# this type of paragrph may be made up of multiple _spans, inline
# word monograms (images) and words with semantic meaning
2009-01-22 18:15:33 +06:00
# and now a new type "span" versus the old "_span"
2010-01-24 18:19:20 +06:00
# plus glyphs used to form starting letter of first word
2010-01-19 18:11:59 +06:00
# need to parse this type line by line
line = start + 1
word_class = ''
2010-01-20 18:13:31 +06:00
# if end is -1 then we must search to end of document
if end == -1 :
2010-01-21 18:14:31 +06:00
end = self.docSize
2010-01-20 18:13:31 +06:00
2010-01-19 18:11:59 +06:00
while (line < end) :
(name, argres) = self.lineinDoc(line)
2009-01-22 18:15:33 +06:00
if name.endswith('span.firstWord') :
2010-01-19 18:11:59 +06:00
first = int(argres)
(name, argres) = self.lineinDoc(line+1)
2009-01-22 18:15:33 +06:00
if not name.endswith('span.lastWord'):
2010-01-19 18:11:59 +06:00
print 'Error: - incorrect _span ordering inside paragraph'
last = int(argres)
for wordnum in xrange(first, last):
result.append(('ocr', wordnum))
line += 1
2010-01-24 18:19:20 +06:00
elif name.endswith('word.firstGlyph') :
first = int(argres)
(name, argres) = self.lineinDoc(line+1)
if not name.endswith('word.lastGlyph'):
print 'Error: - incorrect glyph ordering inside word in paragraph'
last = int(argres)
glyphList = []
for glyphnum in xrange(first, last):
glyphList.append(glyphnum)
num = self.svgcount
self.glyphs_to_image(glyphList)
self.svgcount += 1
result.append(('svg', num))
line += 1
2010-01-19 18:11:59 +06:00
elif name.endswith('word.class'):
(cname, space) = argres.split('-',1)
2010-01-20 18:13:31 +06:00
if space == '' : space = '0'
if (cname == 'spaceafter') and (int(space) > 0) :
2010-01-19 18:11:59 +06:00
word_class = 'sa'
elif name.endswith('word.img.src'):
result.append(('img' + word_class, int(argres)))
word_class = ''
elif name.endswith('word_semantic.firstWord'):
first = int(argres)
(name, argres) = self.lineinDoc(line+1)
if not name.endswith('word_semantic.lastWord'):
print 'Error: - incorrect word_semantic ordering inside paragraph'
last = int(argres)
for wordnum in xrange(first, last):
result.append(('ocr', wordnum))
line += 1
line += 1
return pclass, result
2010-01-21 18:14:31 +06:00
def buildParagraph(self, pclass, pdesc, type, regtype) :
2010-01-17 18:10:35 +06:00
parares = ''
sep =''
2010-01-19 18:11:59 +06:00
2010-01-21 18:14:31 +06:00
classres = ''
if pclass :
classres = ' class="' + pclass + '"'
2010-01-19 18:11:59 +06:00
2009-01-22 18:15:33 +06:00
br_lb = (regtype == 'fixed') or (regtype == 'chapterheading') or (regtype == 'vertical')
2010-01-19 18:11:59 +06:00
2010-01-21 18:14:31 +06:00
handle_links = len(self.link_id) > 0
2010-01-17 18:10:35 +06:00
if (type == 'full') or (type == 'begin') :
2010-01-21 18:14:31 +06:00
parares += '<p' + classres + '>'
2010-01-19 18:11:59 +06:00
2010-01-17 18:10:35 +06:00
if (type == 'end'):
parares += ' '
2010-01-19 18:11:59 +06:00
cnt = len(pdesc)
for j in xrange( 0, cnt) :
(wtype, num) = pdesc[j]
if wtype == 'ocr' :
word = self.ocrtext[num]
sep = ' '
if handle_links:
link = self.link_id[num]
if (link > 0):
title = self.link_title[link-1]
2010-01-20 18:13:31 +06:00
if (title == "") or (parares.rfind(title) < 0):
title='_link_'
2010-01-19 18:11:59 +06:00
ptarget = self.link_page[link-1] - 1
linkhtml = '<a href="#page%04d">' % ptarget
linkhtml += title + '</a>'
pos = parares.rfind(title)
if pos >= 0:
parares = parares[0:pos] + linkhtml + parares[pos+len(title):]
else :
parares += linkhtml
if word == '_link_' : word = ''
elif (link < 0) :
if word == '_link_' : word = ''
if word == '_lb_':
2010-01-21 18:14:31 +06:00
if ((num-1) in self.dehyphen_rootid ) or handle_links:
2010-01-19 18:11:59 +06:00
word = ''
sep = ''
elif br_lb :
word = '<br />\n'
sep = ''
2010-01-17 18:10:35 +06:00
else :
2010-01-19 18:11:59 +06:00
word = '\n'
sep = ''
if num in self.dehyphen_rootid :
word = word[0:-1]
2010-01-17 18:10:35 +06:00
sep = ''
2010-01-19 18:11:59 +06:00
parares += word + sep
elif wtype == 'img' :
2010-01-17 18:10:35 +06:00
sep = ''
2010-01-19 18:11:59 +06:00
parares += '<img src="img/img%04d.jpg" alt="" />' % num
parares += sep
2010-01-17 18:10:35 +06:00
2010-01-19 18:11:59 +06:00
elif wtype == 'imgsa' :
sep = ' '
parares += '<img src="img/img%04d.jpg" alt="" />' % num
parares += sep
2010-01-17 18:10:35 +06:00
2010-01-24 18:19:20 +06:00
elif wtype == 'svg' :
sep = ''
parares += '<img src="img/' + self.id + '_%04d.svg" alt="" />' % num
parares += sep
2010-01-17 18:10:35 +06:00
if len(sep) > 0 : parares = parares[0:-1]
if (type == 'full') or (type == 'end') :
parares += '</p>'
return parares
# walk the document tree collecting the information needed
# to build an html page using the ocrText
def process(self):
htmlpage = ''
2010-01-21 18:14:31 +06:00
# get the ocr text
2010-01-17 18:10:35 +06:00
(pos, argres) = self.findinDoc('info.word.ocrText',0,-1)
if argres : self.ocrtext = argres.split('|')
2010-01-21 18:14:31 +06:00
# get information to dehyphenate the text
2010-01-24 18:19:20 +06:00
self.dehyphen_rootid = self.getData('info.dehyphen.rootID',0,-1)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
# determine if first paragraph is continued from previous page
2010-01-17 18:10:35 +06:00
(pos, self.parastems_stemid) = self.findinDoc('info.paraStems.stemID',0,-1)
2010-01-21 18:14:31 +06:00
first_para_continued = (self.parastems_stemid != None)
# determine if last paragraph is continued onto the next page
2010-01-17 18:10:35 +06:00
(pos, self.paracont_stemid) = self.findinDoc('info.paraCont.stemID',0,-1)
2010-01-21 18:14:31 +06:00
last_para_continued = (self.paracont_stemid != None)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
# collect link ids
2010-01-24 18:19:20 +06:00
self.link_id = self.getData('info.word.link_id',0,-1)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
# collect link destination page numbers
2010-01-24 18:19:20 +06:00
self.link_page = self.getData('info.links.page',0,-1)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
# collect link titles
2010-01-17 18:10:35 +06:00
(pos, argres) = self.findinDoc('info.links.title',0,-1)
if argres :
self.link_title = argres.split('|')
else:
self.link_title.append('')
2010-01-21 18:14:31 +06:00
# get page type
2010-01-17 18:10:35 +06:00
(pos, pagetype) = self.findinDoc('page.type',0,-1)
# generate a list of each region starting point
# each region has one paragraph,, or one image, or one chapterheading
2010-01-21 18:14:31 +06:00
2010-01-17 18:10:35 +06:00
regionList= self.posinDoc('region')
regcnt = len(regionList)
regionList.append(-1)
anchorSet = False
breakSet = False
# process each region tag and convert what you can to html
for j in xrange(regcnt):
2010-01-21 18:14:31 +06:00
2010-01-17 18:10:35 +06:00
start = regionList[j]
end = regionList[j+1]
(pos, regtype) = self.findinDoc('region.type',start,end)
2010-01-21 18:14:31 +06:00
# set anchor for link target on this page
if not anchorSet and not first_para_continued:
2009-01-22 18:15:33 +06:00
htmlpage += '<div style="visibility: hidden; height: 0; width: 0;" id="' + self.id + '" title="pagetype_' + pagetype + '"></div>\n'
2010-01-21 18:14:31 +06:00
anchorSet = True
2010-01-17 18:10:35 +06:00
if regtype == 'graphic' :
(pos, simgsrc) = self.findinDoc('img.src',start,end)
if simgsrc:
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
2010-01-21 18:14:31 +06:00
2010-01-17 18:10:35 +06:00
elif regtype == 'chapterheading' :
2010-01-19 18:11:59 +06:00
(pclass, pdesc) = self.getParaDescription(start,end)
2010-01-17 18:10:35 +06:00
if not breakSet:
htmlpage += '<div style="page-break-after: always;">&nbsp;</div>\n'
breakSet = True
tag = 'h1'
2010-01-21 18:14:31 +06:00
if pclass and (len(pclass) >= 7):
if pclass[3:7] == 'ch1-' : tag = 'h1'
if pclass[3:7] == 'ch2-' : tag = 'h2'
if pclass[3:7] == 'ch3-' : tag = 'h3'
htmlpage += '<' + tag + ' class="' + pclass + '">'
else:
htmlpage += '<' + tag + '>'
2010-01-19 18:11:59 +06:00
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
2010-01-17 18:10:35 +06:00
htmlpage += '</' + tag + '>'
2010-01-21 18:14:31 +06:00
2009-01-22 18:15:33 +06:00
elif (regtype == 'text') or (regtype == 'fixed') or (regtype == 'insert') or (regtype == 'listitem'):
2010-01-17 18:10:35 +06:00
ptype = 'full'
# check to see if this is a continution from the previous page
2010-01-21 18:14:31 +06:00
if first_para_continued :
2010-01-17 18:10:35 +06:00
ptype = 'end'
2010-01-21 18:14:31 +06:00
first_para_continued = False
2010-01-19 18:11:59 +06:00
(pclass, pdesc) = self.getParaDescription(start,end)
2010-01-21 18:14:31 +06:00
if pclass and (len(pclass) >= 6) and (ptype == 'full'):
2010-01-17 18:10:35 +06:00
tag = 'p'
if pclass[3:6] == 'h1-' : tag = 'h4'
if pclass[3:6] == 'h2-' : tag = 'h5'
if pclass[3:6] == 'h3-' : tag = 'h6'
htmlpage += '<' + tag + ' class="' + pclass + '">'
2010-01-19 18:11:59 +06:00
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
2010-01-17 18:10:35 +06:00
htmlpage += '</' + tag + '>'
else :
2010-01-19 18:11:59 +06:00
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
2010-01-17 18:10:35 +06:00
elif (regtype == 'tocentry') :
ptype = 'full'
2010-01-21 18:14:31 +06:00
if first_para_continued :
2010-01-17 18:10:35 +06:00
ptype = 'end'
2010-01-21 18:14:31 +06:00
first_para_continued = False
2010-01-19 18:11:59 +06:00
(pclass, pdesc) = self.getParaDescription(start,end)
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
2010-01-21 18:14:31 +06:00
2009-01-22 18:15:33 +06:00
elif (regtype == 'vertical') :
ptype = 'full'
if first_para_continued :
ptype = 'end'
first_para_continued = False
(pclass, pdesc) = self.getParaDescription(start,end)
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
elif (regtype == 'table') :
2010-01-24 18:19:20 +06:00
# translate first and last word into first and last glyphs
# and generate table as an image and include a link to it
glyphList = []
(pos, sfirst) = self.findinDoc('paragraph.firstWord',start,end)
(pos, slast) = self.findinDoc('paragraph.lastWord',start,end)
firstglyphList = self.getData('word.firstGlyph',0,-1)
gidList = self.getData('info.glyph.glyphID',0,-1)
if (sfirst != None) and (slast != None) :
first = int(sfirst)
last = int(slast)
firstGlyph = firstglyphList[first]
if last < len(firstglyphList):
lastGlyph = firstglyphList[last]
else :
lastGlyph = len(gidList)
for glyphnum in xrange(firstGlyph, lastGlyph):
glyphList.append(glyphnum)
num = self.svgcount
self.glyphs_to_image(glyphList)
self.svgcount += 1
htmlpage += '<div class="graphic"><img src="img/' + self.id + '_%04d.svg" alt="" /></div>' % num
else :
ptype = 'full'
if first_para_continued :
ptype = 'end'
first_para_continued = False
(pclass, pdesc) = self.getParaDescription(start,end)
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
print " "
print "Warning: - Table Conversions are notoriously poor"
print " Strongly recommend taking a screen capture image of the "
print " table in %s.svg and using it to replace this attempt at a table" % self.id
print " "
2009-01-22 18:15:33 +06:00
2010-01-20 18:13:31 +06:00
elif (regtype == 'synth_fcvr.center') or (regtype == 'synth_text.center'):
2010-01-19 18:11:59 +06:00
(pos, simgsrc) = self.findinDoc('img.src',start,end)
if simgsrc:
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
2010-01-17 18:10:35 +06:00
else :
2009-01-22 18:15:33 +06:00
print 'Warning: region type', regtype
2010-01-20 18:13:31 +06:00
(pos, temp) = self.findinDoc('paragraph',start,end)
2009-01-22 18:15:33 +06:00
if pos != -1:
print ' is a "text" region'
2010-01-20 18:13:31 +06:00
regtype = 'fixed'
ptype = 'full'
# check to see if this is a continution from the previous page
2010-01-21 18:14:31 +06:00
if first_para_continued :
2010-01-20 18:13:31 +06:00
ptype = 'end'
2010-01-21 18:14:31 +06:00
first_para_continued = False
2010-01-20 18:13:31 +06:00
(pclass, pdesc) = self.getParaDescription(start,end)
2010-01-21 18:14:31 +06:00
if pclass and (ptype == 'full') and (len(pclass) >= 6):
2010-01-20 18:13:31 +06:00
tag = 'p'
if pclass[3:6] == 'h1-' : tag = 'h4'
if pclass[3:6] == 'h2-' : tag = 'h5'
if pclass[3:6] == 'h3-' : tag = 'h6'
htmlpage += '<' + tag + ' class="' + pclass + '">'
htmlpage += self.buildParagraph(pclass, pdesc, 'middle', regtype)
htmlpage += '</' + tag + '>'
else :
htmlpage += self.buildParagraph(pclass, pdesc, ptype, regtype)
else :
2009-01-22 18:15:33 +06:00
print ' is a "graphic" region'
2010-01-20 18:13:31 +06:00
(pos, simgsrc) = self.findinDoc('img.src',start,end)
if simgsrc:
htmlpage += '<div class="graphic"><img src="img/img%04d.jpg" alt="" /></div>' % int(simgsrc)
2010-01-17 18:10:35 +06:00
2010-01-21 18:14:31 +06:00
if last_para_continued :
2010-01-17 18:10:35 +06:00
if htmlpage[-4:] == '</p>':
2010-01-21 18:14:31 +06:00
htmlpage = htmlpage[0:-4]
last_para_continued = False
2010-01-17 18:10:35 +06:00
return htmlpage
2010-01-24 18:19:20 +06:00
def convert2HTML(flatxml, classlst, fileid, bookDir):
2010-01-17 18:10:35 +06:00
# create a document parser
2010-01-24 18:19:20 +06:00
dp = DocParser(flatxml, classlst, fileid, bookDir)
2010-01-17 18:10:35 +06:00
htmlpage = dp.process()
return htmlpage