|
|
| version 1.175, 2011/06/14 09:57:11 | version 1.175.2.9, 2011/07/27 19:09:44 |
|---|---|
| Line 1 | Line 1 |
| from OFS.Folder import Folder | from OFS.Folder import Folder |
| from Products.PageTemplates.ZopePageTemplate import ZopePageTemplate | from Products.PageTemplates.ZopePageTemplate import ZopePageTemplate |
| from Products.PageTemplates.PageTemplateFile import PageTemplateFile | from Products.PageTemplates.PageTemplateFile import PageTemplateFile |
| from AccessControl import ClassSecurityInfo | from AccessControl import ClassSecurityInfo |
| from AccessControl import getSecurityManager | from AccessControl import getSecurityManager |
| from Globals import package_home | from Globals import package_home |
| from Products.zogiLib.zogiLib import browserCheck | |
| from Ft.Xml import EMPTY_NAMESPACE, Parse | #from Ft.Xml import EMPTY_NAMESPACE, Parse |
| import Ft.Xml.Domlette | #import Ft.Xml.Domlette |
| import xml.etree.ElementTree as ET | |
| import os.path | import os.path |
| import sys | import sys |
| import urllib | import urllib |
| import urllib2 | |
| import logging | import logging |
| import math | import math |
| import urlparse | import urlparse |
| import cStringIO | |
| import re | import re |
| import string | import string |
| from SrvTxtUtils import getInt, getText, getHttpData | |
| def logger(txt,method,txt2): | def logger(txt,method,txt2): |
| """logging""" | """logging""" |
| logging.info(txt+ txt2) | logging.info(txt+ txt2) |
| def getInt(number, default=0): | |
| """returns always an int (0 in case of problems)""" | |
| try: | |
| return int(number) | |
| except: | |
| return int(default) | |
| def getTextFromNode(nodename): | |
| """get the cdata content of a node""" | |
| if nodename is None: | |
| return "" | |
| nodelist=nodename.childNodes | |
| rc = "" | |
| for node in nodelist: | |
| if node.nodeType == node.TEXT_NODE: | |
| rc = rc + node.data | |
| return rc | |
| def serializeNode(node, encoding="utf-8"): | def serializeNode(node, encoding="utf-8"): |
| """returns a string containing node as XML""" | """returns a string containing node as XML""" |
| stream = cStringIO.StringIO() | s = ET.tostring(node) |
| #logging.debug("BUF: %s"%(stream)) | |
| Ft.Xml.Domlette.Print(node, stream=stream, encoding=encoding) | # 4Suite: |
| s = stream.getvalue() | # stream = cStringIO.StringIO() |
| #logging.debug("BUF: %s"%(s)) | # Ft.Xml.Domlette.Print(node, stream=stream, encoding=encoding) |
| stream.close() | # s = stream.getvalue() |
| # stream.close() | |
| return s | return s |
| def browserCheck(self): | def browserCheck(self): |
| Line 114 def browserCheck(self): | Line 98 def browserCheck(self): |
| return bt | return bt |
| def getParentDir(path): | def getParentDir(path): |
| """returns pathname shortened by one""" | """returns pathname shortened by one""" |
| return '/'.join(path.split('/')[0:-1]) | return '/'.join(path.split('/')[0:-1]) |
| def normalizeBibField(bt, underscore=True): | |
| """returns normalised bib type for looking up mappings""" | |
| bt = bt.strip().replace(' ', '-').lower() | |
| if underscore: | |
| bt = bt.replace('_', '-') | |
| def getHttpData(url, data=None, num_tries=3, timeout=10): | return bt |
| """returns result from url+data HTTP request""" | |
| # we do GET (by appending data to url) | def getBibdataFromDom(dom): |
| if isinstance(data, str) or isinstance(data, unicode): | """returns dict with all elements from bib-tag""" |
| # if data is string then append | bibinfo = {} |
| url = "%s?%s"%(url,data) | bib = dom.find(".//meta/bib") |
| elif isinstance(data, dict) or isinstance(data, list) or isinstance(data, tuple): | if bib is not None: |
| # urlencode | # put type in @type |
| url = "%s?%s"%(url,urllib.urlencode(data)) | type = bib.get('type') |
| bibinfo['@type'] = normalizeBibField(type) | |
| response = None | # put all subelements in dict |
| errmsg = None | for e in bib: |
| for cnt in range(num_tries): | bibinfo[normalizeBibField(e.tag)] = getText(e) |
| try: | |
| logging.debug("getHttpData(#%s %ss) url=%s"%(cnt+1,timeout,url)) | return bibinfo |
| if sys.version_info < (2, 6): | |
| # set timeout on socket -- ugly :-( | |
| import socket | |
| socket.setdefaulttimeout(float(timeout)) | |
| response = urllib2.urlopen(url) | |
| else: | |
| response = urllib2.urlopen(url,timeout=float(timeout)) | |
| # check result? | |
| break | |
| except urllib2.HTTPError, e: | |
| logging.error("getHttpData: HTTP error(%s): %s"%(e.code,e)) | |
| errmsg = str(e) | |
| # stop trying | |
| break | |
| except urllib2.URLError, e: | |
| logging.error("getHttpData: URLLIB error(%s): %s"%(e.reason,e)) | |
| errmsg = str(e) | |
| # stop trying | |
| #break | |
| if response is not None: | |
| data = response.read() | |
| response.close() | |
| return data | |
| raise IOError("ERROR fetching HTTP data from %s: %s"%(url,errmsg)) | |
| #return None | |
| ## | ## |
| ## documentViewer class | ## documentViewer class |
| Line 290 class documentViewer(Folder): | Line 252 class documentViewer(Folder): |
| ''' | ''' |
| logging.debug("HHHHHHHHHHHHHH:load the rss") | logging.debug("HHHHHHHHHHHHHH:load the rss") |
| logger("documentViewer (index)", logging.INFO, "mode: %s url:%s start:%s pn:%s"%(mode,url,start,pn)) | logging.debug("documentViewer (index) mode: %s url:%s start:%s pn:%s"%(mode,url,start,pn)) |
| if not hasattr(self, 'template'): | if not hasattr(self, 'template'): |
| # create template folder if it doesn't exist | # create template folder if it doesn't exist |
| Line 341 class documentViewer(Folder): | Line 303 class documentViewer(Folder): |
| # get table of contents | # get table of contents |
| docinfo = self.getToc(mode=tocMode, docinfo=docinfo) | docinfo = self.getToc(mode=tocMode, docinfo=docinfo) |
| if viewMode=="auto": # automodus gewaehlt | # auto viewMode: text_dict if text else images |
| if docinfo.has_key('textURL') or docinfo.get('textURLPath',None): #texturl gesetzt und textViewer konfiguriert | if viewMode=="auto": |
| if docinfo.get('textURL', None) or docinfo.get('textURLPath', None): | |
| #texturl gesetzt und textViewer konfiguriert | |
| viewMode="text_dict" | viewMode="text_dict" |
| else: | else: |
| viewMode="images" | viewMode="images" |
| pageinfo = self.getPageinfo(start=start,current=pn, docinfo=docinfo,viewMode=viewMode,tocMode=tocMode) | pageinfo = self.getPageinfo(start=start,current=pn, docinfo=docinfo,viewMode=viewMode,tocMode=tocMode) |
| if (docinfo.get('textURLPath',None)): | if viewMode != 'images' and docinfo.get('textURLPath', None): |
| page = self.getTextPage(docinfo=docinfo, pageinfo=pageinfo) | # get full text page |
| page = self.getTextPage(mode=viewMode, pn=pn, docinfo=docinfo, pageinfo=pageinfo) | |
| pageinfo['textPage'] = page | pageinfo['textPage'] = page |
| tt = getattr(self, 'template') | |
| pt = getattr(tt, 'viewer_main') | # get template /template/viewer_main |
| pt = getattr(self.template, 'viewer_main') | |
| # and execute with parameters | |
| return pt(docinfo=docinfo,pageinfo=pageinfo,viewMode=viewMode,mk=self.generateMarks(mk)) | return pt(docinfo=docinfo,pageinfo=pageinfo,viewMode=viewMode,mk=self.generateMarks(mk)) |
| def generateMarks(self,mk): | def generateMarks(self,mk): |
| Line 485 class documentViewer(Folder): | Line 452 class documentViewer(Folder): |
| docinfo = {} | docinfo = {} |
| for x in range(cut): | for x in range(cut): |
| path=getParentDir(path) | path=getParentDir(path) |
| infoUrl=self.digilibBaseUrl+"/dirInfo-xml.jsp?mo=dir&fn="+path | infoUrl=self.digilibBaseUrl+"/dirInfo-xml.jsp?mo=dir&fn="+path |
| Line 496 class documentViewer(Folder): | Line 462 class documentViewer(Folder): |
| if txt is None: | if txt is None: |
| raise IOError("Unable to get dir-info from %s"%(infoUrl)) | raise IOError("Unable to get dir-info from %s"%(infoUrl)) |
| dom = Parse(txt) | dom = ET.fromstring(txt) |
| sizes=dom.xpath("//dir/size") | #dom = Parse(txt) |
| logging.debug("documentViewer (getparamfromdigilib) dirInfo:size"%sizes) | size=getText(dom.find("size")) |
| #sizes=dom.xpath("//dir/size") | |
| logging.debug("documentViewer (getparamfromdigilib) dirInfo:size=%s"%size) | |
| if sizes: | if size: |
| docinfo['numPages'] = int(getTextFromNode(sizes[0])) | docinfo['numPages'] = int(size) |
| else: | else: |
| docinfo['numPages'] = 0 | docinfo['numPages'] = 0 |
| Line 546 class documentViewer(Folder): | Line 514 class documentViewer(Folder): |
| if txt is None: | if txt is None: |
| raise IOError("Unable to read index meta from %s"%(url)) | raise IOError("Unable to read index meta from %s"%(url)) |
| dom = Parse(txt) | dom = ET.fromstring(txt) |
| #dom = Parse(txt) | |
| return dom | return dom |
| def getPresentationInfoXML(self, url): | def getPresentationInfoXML(self, url): |
| Line 565 class documentViewer(Folder): | Line 534 class documentViewer(Folder): |
| if txt is None: | if txt is None: |
| raise IOError("Unable to read infoXMLfrom %s"%(url)) | raise IOError("Unable to read infoXMLfrom %s"%(url)) |
| dom = Parse(txt) | dom = ET.fromstring(txt) |
| #dom = Parse(txt) | |
| return dom | return dom |
| Line 583 class documentViewer(Folder): | Line 553 class documentViewer(Folder): |
| path=getParentDir(path) | path=getParentDir(path) |
| dom = self.getDomFromIndexMeta(path) | dom = self.getDomFromIndexMeta(path) |
| acctype = dom.xpath("//access-conditions/access/@type") | acc = dom.find(".//access-conditions/access") |
| if acctype and (len(acctype)>0): | if acc is not None: |
| access=acctype[0].value | acctype = acc.get('type') |
| #acctype = dom.xpath("//access-conditions/access/@type") | |
| if acctype: | |
| access=acctype | |
| if access in ['group', 'institution']: | if access in ['group', 'institution']: |
| access = getTextFromNode(dom.xpath("//access-conditions/access/name")[0]).lower() | access = dom.find(".//access-conditions/access/name").text.lower() |
| docinfo['accessType'] = access | docinfo['accessType'] = access |
| return docinfo | return docinfo |
| Line 608 class documentViewer(Folder): | Line 581 class documentViewer(Folder): |
| docinfo['indexMetaPath']=self.getIndexMetaPath(path); | docinfo['indexMetaPath']=self.getIndexMetaPath(path); |
| logging.debug("documentViewer (getbibinfofromindexmeta cutted) path: %s"%(path)) | logging.debug("documentViewer (getbibinfofromindexmeta cutted) path: %s"%(path)) |
| # put in all raw bib fields as dict "bib" | # try to get MetaDataFolder |
| bib = dom.xpath("//bib/*") | metadata = getattr(self, 'metadata') |
| if bib and len(bib)>0: | if metadata is not None: |
| bibinfo = {} | # put all raw bib fields in dict "bib" |
| for e in bib: | bib = metadata.getBibdataFromDom(dom) |
| bibinfo[e.localName] = getTextFromNode(e) | docinfo['bib'] = bib |
| docinfo['bib'] = bibinfo | bibtype = bib.get('@type', None) |
| # extract some fields (author, title, year) according to their mapping | |
| metaData=self.metadata.main.meta.bib | |
| bibtype=dom.xpath("//bib/@type") | |
| if bibtype and (len(bibtype)>0): | |
| bibtype=bibtype[0].value | |
| else: | |
| bibtype="generic" | |
| bibtype=bibtype.replace("-"," ") # wrong typesiin index meta "-" instead of " " (not wrong! ROC) | |
| docinfo['bib_type'] = bibtype | docinfo['bib_type'] = bibtype |
| bibmap=metaData.generateMappingForType(bibtype) | # also store DC metadata for convenience |
| logging.debug("documentViewer (getbibinfofromindexmeta) bibmap:"+repr(bibmap)) | dc = metadata.getDCMappedData(bib) |
| logging.debug("documentViewer (getbibinfofromindexmeta) bibtype:"+repr(bibtype)) | docinfo['creator'] = dc.get('creator',None) |
| # if there is no mapping bibmap is empty (mapping sometimes has empty fields) | docinfo['title'] = dc.get('title',None) |
| if len(bibmap) > 0 and len(bibmap['author'][0]) > 0: | docinfo['date'] = dc.get('date',None) |
| try: | else: |
| docinfo['author']=getTextFromNode(dom.xpath("//bib/%s"%bibmap['author'][0])[0]) | logging.error("MetaDataFolder 'metadata' not found!") |
| except: pass | #TODO: remove |
| try: | bib = getBibdataFromDom(dom) |
| docinfo['title']=getTextFromNode(dom.xpath("//bib/%s"%bibmap['title'][0])[0]) | |
| except: pass | |
| try: | |
| docinfo['year']=getTextFromNode(dom.xpath("//bib/%s"%bibmap['year'][0])[0]) | |
| except: pass | |
| logging.debug("documentViewer (getbibinfofromindexmeta) using mapping for %s"%bibtype) | |
| try: | |
| docinfo['lang']=getTextFromNode(dom.xpath("//bib/lang")[0]) | |
| except: | |
| docinfo['lang']='' | |
| try: | |
| docinfo['city']=getTextFromNode(dom.xpath("//bib/city")[0]) | |
| except: | |
| docinfo['city']='' | |
| try: | |
| docinfo['number_of_pages']=getTextFromNode(dom.xpath("//bib/number_of_pages")[0]) | |
| except: | |
| docinfo['number_of_pages']='' | |
| try: | |
| docinfo['series_volume']=getTextFromNode(dom.xpath("//bib/series_volume")[0]) | |
| except: | |
| docinfo['series_volume']='' | |
| try: | |
| docinfo['number_of_volumes']=getTextFromNode(dom.xpath("//bib/number_of_volumes")[0]) | |
| except: | |
| docinfo['number_of_volumes']='' | |
| try: | |
| docinfo['translator']=getTextFromNode(dom.xpath("//bib/translator")[0]) | |
| except: | |
| docinfo['translator']='' | |
| try: | |
| docinfo['edition']=getTextFromNode(dom.xpath("//bib/edition")[0]) | |
| except: | |
| docinfo['edition']='' | |
| try: | |
| docinfo['series_author']=getTextFromNode(dom.xpath("//bib/series_author")[0]) | |
| except: | |
| docinfo['series_author']='' | |
| try: | |
| docinfo['publisher']=getTextFromNode(dom.xpath("//bib/publisher")[0]) | |
| except: | |
| docinfo['publisher']='' | |
| try: | |
| docinfo['series_title']=getTextFromNode(dom.xpath("//bib/series_title")[0]) | |
| except: | |
| docinfo['series_title']='' | |
| try: | |
| docinfo['isbn_issn']=getTextFromNode(dom.xpath("//bib/isbn_issn")[0]) | |
| except: | |
| docinfo['isbn_issn']='' | |
| return docinfo | return docinfo |
| # TODO: is this needed? | |
| def getNameFromIndexMeta(self,path,docinfo=None,dom=None,cut=0): | def getNameFromIndexMeta(self,path,docinfo=None,dom=None,cut=0): |
| """gets name info from the index.meta file at path or given by dom""" | """gets name info from the index.meta file at path or given by dom""" |
| if docinfo is None: | if docinfo is None: |
| Line 698 class documentViewer(Folder): | Line 612 class documentViewer(Folder): |
| path=getParentDir(path) | path=getParentDir(path) |
| dom = self.getDomFromIndexMeta(path) | dom = self.getDomFromIndexMeta(path) |
| docinfo['name']=getTextFromNode(dom.xpath("/resource/name")[0]) | docinfo['name']=getText(dom.find("name")) |
| logging.debug("documentViewer docinfo[name] %s"%docinfo['name']) | logging.debug("documentViewer docinfo[name] %s"%docinfo['name']) |
| return docinfo | return docinfo |
| Line 715 class documentViewer(Folder): | Line 629 class documentViewer(Folder): |
| archivePath = None | archivePath = None |
| archiveName = None | archiveName = None |
| archiveNames = dom.xpath("//resource/name") | archiveName = getText(dom.find("name")) |
| if archiveNames and (len(archiveNames) > 0): | if not archiveName: |
| archiveName = getTextFromNode(archiveNames[0]) | |
| else: | |
| logging.warning("documentViewer (getdocinfofromtexttool) resource/name missing in: %s" % (url)) | logging.warning("documentViewer (getdocinfofromtexttool) resource/name missing in: %s" % (url)) |
| archivePaths = dom.xpath("//resource/archive-path") | archivePath = getText(dom.find("archive-path")) |
| if archivePaths and (len(archivePaths) > 0): | if archivePath: |
| archivePath = getTextFromNode(archivePaths[0]) | |
| # clean up archive path | # clean up archive path |
| if archivePath[0] != '/': | if archivePath[0] != '/': |
| archivePath = '/' + archivePath | archivePath = '/' + archivePath |
| Line 739 class documentViewer(Folder): | Line 650 class documentViewer(Folder): |
| # we balk without archive-path | # we balk without archive-path |
| raise IOError("Missing archive-path (for text-tool) in %s" % (url)) | raise IOError("Missing archive-path (for text-tool) in %s" % (url)) |
| imageDirs = dom.xpath("//texttool/image") | imageDir = getText(dom.find(".//texttool/image")) |
| if imageDirs and (len(imageDirs) > 0): | |
| imageDir = getTextFromNode(imageDirs[0]) | |
| else: | if not imageDir: |
| # we balk with no image tag / not necessary anymore because textmode is now standard | # we balk with no image tag / not necessary anymore because textmode is now standard |
| #raise IOError("No text-tool info in %s"%(url)) | #raise IOError("No text-tool info in %s"%(url)) |
| imageDir = "" | imageDir = "" |
| Line 760 class documentViewer(Folder): | Line 669 class documentViewer(Folder): |
| docinfo['imageURL'] = self.digilibBaseUrl + "/servlet/Scaler?fn=" + imageDir | docinfo['imageURL'] = self.digilibBaseUrl + "/servlet/Scaler?fn=" + imageDir |
| viewerUrls = dom.xpath("//texttool/digiliburlprefix") | viewerUrl = getText(dom.find(".//texttool/digiliburlprefix")) |
| if viewerUrls and (len(viewerUrls) > 0): | if viewerUrl: |
| viewerUrl = getTextFromNode(viewerUrls[0]) | |
| docinfo['viewerURL'] = viewerUrl | docinfo['viewerURL'] = viewerUrl |
| # old style text URL | # old style text URL |
| textUrls = dom.xpath("//texttool/text") | textUrl = getText(dom.find(".//texttool/text")) |
| if textUrls and (len(textUrls) > 0): | if textUrl: |
| textUrl = getTextFromNode(textUrls[0]) | |
| if urlparse.urlparse(textUrl)[0] == "": #keine url | if urlparse.urlparse(textUrl)[0] == "": #keine url |
| textUrl = os.path.join(archivePath, textUrl) | textUrl = os.path.join(archivePath, textUrl) |
| # fix URLs starting with /mpiwg/online | # fix URLs starting with /mpiwg/online |
| Line 778 class documentViewer(Folder): | Line 685 class documentViewer(Folder): |
| docinfo['textURL'] = textUrl | docinfo['textURL'] = textUrl |
| # new style text-url-path | # new style text-url-path |
| textUrls = dom.xpath("//texttool/text-url-path") | textUrl = getText(dom.find(".//texttool/text-url-path")) |
| if textUrls and (len(textUrls) > 0): | if textUrl: |
| textUrl = getTextFromNode(textUrls[0]) | |
| docinfo['textURLPath'] = textUrl | docinfo['textURLPath'] = textUrl |
| textUrlkurz = string.split(textUrl, ".")[0] | textUrlkurz = string.split(textUrl, ".")[0] |
| docinfo['textURLPathkurz'] = textUrlkurz | docinfo['textURLPathkurz'] = textUrlkurz |
| Line 789 class documentViewer(Folder): | Line 695 class documentViewer(Folder): |
| #docinfo = self.getNumTextPages(docinfo) | #docinfo = self.getNumTextPages(docinfo) |
| presentationUrls = dom.xpath("//texttool/presentation") | presentationUrl = getText(dom.find(".//texttool/presentation")) |
| docinfo = self.getBibinfoFromIndexMeta(url, docinfo=docinfo, dom=dom) # get info von bib tag | docinfo = self.getBibinfoFromIndexMeta(url, docinfo=docinfo, dom=dom) # get info von bib tag |
| # TODO: is this needed here? | |
| docinfo = self.getNameFromIndexMeta(url, docinfo=docinfo, dom=dom) | docinfo = self.getNameFromIndexMeta(url, docinfo=docinfo, dom=dom) |
| if presentationUrls and (len(presentationUrls) > 0): # ueberschreibe diese durch presentation informationen | if presentationUrl: # ueberschreibe diese durch presentation informationen |
| # presentation url ergiebt sich ersetzen von index.meta in der url der fuer die Metadaten | # presentation url ergiebt sich ersetzen von index.meta in der url der fuer die Metadaten |
| # durch den relativen Pfad auf die presentation infos | # durch den relativen Pfad auf die presentation infos |
| presentationPath = getTextFromNode(presentationUrls[0]) | presentationPath = presentationUrl |
| if url.endswith("index.meta"): | if url.endswith("index.meta"): |
| presentationUrl = url.replace('index.meta', presentationPath) | presentationUrl = url.replace('index.meta', presentationPath) |
| else: | else: |
| Line 814 class documentViewer(Folder): | Line 721 class documentViewer(Folder): |
| """gets the bibliographical information from the preseantion entry in texttools | """gets the bibliographical information from the preseantion entry in texttools |
| """ | """ |
| dom=self.getPresentationInfoXML(url) | dom=self.getPresentationInfoXML(url) |
| try: | docinfo['author']=getText(dom.find(".//author")) |
| docinfo['author']=getTextFromNode(dom.xpath("//author")[0]) | docinfo['title']=getText(dom.find(".//title")) |
| except: | docinfo['year']=getText(dom.find(".//date")) |
| pass | |
| try: | |
| docinfo['title']=getTextFromNode(dom.xpath("//title")[0]) | |
| except: | |
| pass | |
| try: | |
| docinfo['year']=getTextFromNode(dom.xpath("//date")[0]) | |
| except: | |
| pass | |
| return docinfo | return docinfo |
| def getDocinfoFromImagePath(self,path,docinfo=None,cut=0): | def getDocinfoFromImagePath(self,path,docinfo=None,cut=0): |
| Line 844 class documentViewer(Folder): | Line 742 class documentViewer(Folder): |
| imageUrl=self.digilibBaseUrl+"/servlet/Scaler?fn="+path | imageUrl=self.digilibBaseUrl+"/servlet/Scaler?fn="+path |
| docinfo['imageURL'] = imageUrl | docinfo['imageURL'] = imageUrl |
| #TODO: use getDocinfoFromIndexMeta | |
| #path ist the path to the images it assumes that the index.meta file is one level higher. | #path ist the path to the images it assumes that the index.meta file is one level higher. |
| docinfo = self.getBibinfoFromIndexMeta(pathorig,docinfo=docinfo,cut=cut+1) | docinfo = self.getBibinfoFromIndexMeta(pathorig,docinfo=docinfo,cut=cut+1) |
| docinfo = self.getAuthinfoFromIndexMeta(pathorig,docinfo=docinfo,cut=cut+1) | docinfo = self.getAuthinfoFromIndexMeta(pathorig,docinfo=docinfo,cut=cut+1) |
| Line 858 class documentViewer(Folder): | Line 757 class documentViewer(Folder): |
| docinfo = self.REQUEST.SESSION['docinfo'] | docinfo = self.REQUEST.SESSION['docinfo'] |
| # check if its still current | # check if its still current |
| if docinfo is not None and docinfo.get('mode') == mode and docinfo.get('url') == url: | if docinfo is not None and docinfo.get('mode') == mode and docinfo.get('url') == url: |
| logging.debug("documentViewer (getdocinfo) docinfo in session: %s"%docinfo) | logging.debug("documentViewer (getdocinfo) docinfo in session. keys=%s"%docinfo.keys()) |
| return docinfo | return docinfo |
| # new docinfo | # new docinfo |
| docinfo = {'mode': mode, 'url': url} | docinfo = {'mode': mode, 'url': url} |
| if mode=="texttool": #index.meta with texttool information | # add self url |
| docinfo['viewerUrl'] = self.getDocumentViewerURL() | |
| if mode=="texttool": | |
| # index.meta with texttool information | |
| docinfo = self.getDocinfoFromTextTool(url, docinfo=docinfo) | docinfo = self.getDocinfoFromTextTool(url, docinfo=docinfo) |
| elif mode=="imagepath": | elif mode=="imagepath": |
| # folder with images, index.meta optional | |
| docinfo = self.getDocinfoFromImagePath(url, docinfo=docinfo) | docinfo = self.getDocinfoFromImagePath(url, docinfo=docinfo) |
| elif mode=="filepath": | elif mode=="filepath": |
| # filename | |
| docinfo = self.getDocinfoFromImagePath(url, docinfo=docinfo,cut=1) | docinfo = self.getDocinfoFromImagePath(url, docinfo=docinfo,cut=1) |
| else: | else: |
| logging.error("documentViewer (getdocinfo) unknown mode: %s!"%mode) | logging.error("documentViewer (getdocinfo) unknown mode: %s!"%mode) |
| raise ValueError("Unknown mode %s! Has to be one of 'texttool','imagepath','filepath'."%(mode)) | raise ValueError("Unknown mode %s! Has to be one of 'texttool','imagepath','filepath'."%(mode)) |
| # FIXME: fake texturlpath | logging.debug("documentViewer (getdocinfo) docinfo: keys=%s"%docinfo.keys()) |
| if not docinfo.has_key('textURLPath'): | #logging.debug("documentViewer (getdocinfo) docinfo: %s"%docinfo) |
| docinfo['textURLPath'] = None | # store in session |
| logging.debug("documentViewer (getdocinfo) docinfo: %s"%docinfo) | |
| #logging.debug("documentViewer (getdocinfo) docinfo: %s"%) | |
| self.REQUEST.SESSION['docinfo'] = docinfo | self.REQUEST.SESSION['docinfo'] = docinfo |
| return docinfo | return docinfo |
| Line 893 class documentViewer(Folder): | Line 795 class documentViewer(Folder): |
| pageinfo['cols'] = cols | pageinfo['cols'] = cols |
| grpsize = cols * rows | grpsize = cols * rows |
| pageinfo['groupsize'] = grpsize | pageinfo['groupsize'] = grpsize |
| # what does this do? | |
| start = getInt(start, default=(math.ceil(float(current)/float(grpsize))*grpsize-(grpsize-1))) | start = getInt(start, default=(math.ceil(float(current)/float(grpsize))*grpsize-(grpsize-1))) |
| # int(current / grpsize) * grpsize +1)) | # int(current / grpsize) * grpsize +1)) |
| pageinfo['start'] = start | pageinfo['start'] = start |
| Line 903 class documentViewer(Folder): | Line 806 class documentViewer(Folder): |
| pageinfo['numgroups'] = int(np / grpsize) | pageinfo['numgroups'] = int(np / grpsize) |
| if np % grpsize > 0: | if np % grpsize > 0: |
| pageinfo['numgroups'] += 1 | pageinfo['numgroups'] += 1 |
| pageinfo['viewMode'] = viewMode | pageinfo['viewMode'] = viewMode |
| pageinfo['tocMode'] = tocMode | pageinfo['tocMode'] = tocMode |
| pageinfo['characterNormalization'] = self.REQUEST.get('characterNormalization','reg') | pageinfo['characterNormalization'] = self.REQUEST.get('characterNormalization','reg') |
| Line 915 class documentViewer(Folder): | Line 819 class documentViewer(Folder): |
| pageinfo['tocPageSize'] = self.REQUEST.get('tocPageSize', '30') | pageinfo['tocPageSize'] = self.REQUEST.get('tocPageSize', '30') |
| pageinfo['queryPageSize'] =self.REQUEST.get('queryPageSize', '10') | pageinfo['queryPageSize'] =self.REQUEST.get('queryPageSize', '10') |
| pageinfo['tocPN'] = self.REQUEST.get('tocPN', '1') | pageinfo['tocPN'] = self.REQUEST.get('tocPN', '1') |
| # WTF?: | |
| toc = int (pageinfo['tocPN']) | toc = int (pageinfo['tocPN']) |
| pageinfo['textPages'] =int (toc) | pageinfo['textPages'] =int (toc) |
| # What does this do? | |
| if 'tocSize_%s'%tocMode in docinfo: | if 'tocSize_%s'%tocMode in docinfo: |
| tocSize = int(docinfo['tocSize_%s'%tocMode]) | tocSize = int(docinfo['tocSize_%s'%tocMode]) |
| tocPageSize = int(pageinfo['tocPageSize']) | tocPageSize = int(pageinfo['tocPageSize']) |
| Line 926 class documentViewer(Folder): | Line 832 class documentViewer(Folder): |
| tocPages=tocSize/tocPageSize+1 | tocPages=tocSize/tocPageSize+1 |
| else: | else: |
| tocPages=tocSize/tocPageSize | tocPages=tocSize/tocPageSize |
| pageinfo['tocPN'] = min (tocPages,toc) | pageinfo['tocPN'] = min (tocPages,toc) |
| pageinfo['searchPN'] =self.REQUEST.get('searchPN','1') | pageinfo['searchPN'] =self.REQUEST.get('searchPN','1') |
| pageinfo['sn'] =self.REQUEST.get('sn','') | pageinfo['sn'] =self.REQUEST.get('sn','') |
| return pageinfo | return pageinfo |
| def changeDocumentViewer(self,title="",digilibBaseUrl=None,thumbrows=2,thumbcols=5,authgroups='mpiwg',RESPONSE=None): | def changeDocumentViewer(self,title="",digilibBaseUrl=None,thumbrows=2,thumbcols=5,authgroups='mpiwg',RESPONSE=None): |
| """init document viewer""" | """init document viewer""" |
| self.title=title | self.title=title |