--- documentViewer/documentViewer.py 2010/06/18 19:41:28 1.65.2.2 +++ documentViewer/documentViewer.py 2010/06/16 16:38:17 1.69.2.2 @@ -2,20 +2,32 @@ from OFS.Folder import Folder from Products.PageTemplates.ZopePageTemplate import ZopePageTemplate from Products.PageTemplates.PageTemplateFile import PageTemplateFile +from Products.PythonScripts.standard import url_quote from AccessControl import ClassSecurityInfo from AccessControl import getSecurityManager from Globals import package_home +from Ft.Xml.Domlette import NonvalidatingReader +from Ft.Xml.Domlette import PrettyPrint, Print from Ft.Xml import EMPTY_NAMESPACE, Parse -import Ft.Xml.Domlette + +from xml.dom.minidom import parse, parseString + +from extraFunction import * + + +import Ft.Xml.XPath +import cStringIO +import xmlrpclib import os.path import sys +import cgi import urllib -import urllib2 import logging import math + import urlparse -import cStringIO +from types import * def logger(txt,method,txt2): """logging""" @@ -43,7 +55,7 @@ def getTextFromNode(nodename): def serializeNode(node, encoding='utf-8'): """returns a string containing node as XML""" buf = cStringIO.StringIO() - Ft.Xml.Domlette.Print(node, stream=buf, encoding=encoding) + Print(node, stream=buf, encoding=encoding) s = buf.getvalue() buf.close() return s @@ -54,56 +66,23 @@ def getParentDir(path): return '/'.join(path.split('/')[0:-1]) -def getHttpData(url, data=None, num_tries=3, timeout=10): - """returns result from url+data HTTP request""" - # we do GET (by appending data to url) - if isinstance(data, str) or isinstance(data, unicode): - # if data is string then append - url = "%s?%s"%(url,data) - elif isinstance(data, dict) or isinstance(data, list) or isinstance(data, tuple): - # urlencode - url = "%s?%s"%(url,urllib.urlencode(data)) - - response = None - errmsg = None - for cnt in range(num_tries): - try: - logging.debug("getHttpData(#%s %ss) url=%s"%(cnt+1,timeout,url)) - if sys.version_info < (2, 6): - # set timeout on socket -- ugly :-( - import socket - socket.setdefaulttimeout(float(timeout)) - response = urllib2.urlopen(url) - else: - response = urllib2.urlopen(url,timeout=float(timeout)) - # check result? - break - except urllib2.HTTPError, e: - logging.error("getHttpData: HTTP error(%s): %s"%(e.code,e)) - errmsg = str(e) - # stop trying - break - except urllib2.URLError, e: - logging.error("getHttpData: URLLIB error(%s): %s"%(e.reason,e)) - errmsg = str(e) - # stop trying - #break - - if response is not None: - data = response.read() - response.close() - return data - - raise IOError("ERROR fetching HTTP data from %s: %s"%(url,errmsg)) - #return None +import socket +def urlopen(url,timeout=2): + """urlopen mit timeout""" + socket.setdefaulttimeout(timeout) + ret=urllib.urlopen(url) + socket.setdefaulttimeout(5) + return ret ## ## documentViewer class ## -class documentViewer(Folder): +class documentViewer(Folder, extraFunction): """document viewer""" + #textViewerUrl="http://127.0.0.1:8080/HFQP/testXSLT/getPage?" + meta_type="Document viewer" security=ClassSecurityInfo() @@ -144,12 +123,12 @@ class documentViewer(Folder): #self['template'] = templateFolder # Zope-2.12 style self._setObject('template',templateFolder) # old style try: - import MpdlXmlTextServer - textServer = MpdlXmlTextServer.MpdlXmlTextServer(id='fulltextclient',serverName=textServerName) + from Products.XMLRpcTools.XMLRpcTools import XMLRpcServerProxy + xmlRpcClient = XMLRpcServerProxy(id='fulltextclient', serverUrl=textServerName, use_xmlrpc=False) #templateFolder['fulltextclient'] = xmlRpcClient - templateFolder._setObject('fulltextclient',textServer) + templateFolder._setObject('fulltextclient',xmlRpcClient) except Exception, e: - logging.error("Unable to create MpdlXmlTextServer for fulltextclient: "+str(e)) + logging.error("Unable to create XMLRpcTools for fulltextclient: "+str(e)) try: from Products.zogiLib.zogiLib import zogiLib zogilib = zogiLib(id="zogilib", title="zogilib for docuviewer", dlServerURL=imageScalerUrl, layout="book") @@ -158,41 +137,7 @@ class documentViewer(Folder): except Exception, e: logging.error("Unable to create zogiLib for zogilib: "+str(e)) - - # proxy text server methods to fulltextclient - def getTextPage(self, **args): - """get page""" - return self.template.fulltextclient.getTextPage(**args) - - def getQuery(self, **args): - """get query""" - return self.template.fulltextclient.getQuery(**args) - - def getSearch(self, **args): - """get search""" - return self.template.fulltextclient.getSearch(**args) - - def getNumPages(self, docinfo): - """get numpages""" - return self.template.fulltextclient.getNumPages(docinfo) - - def getTranslate(self, **args): - """get translate""" - return self.template.fulltextclient.getTranslate(**args) - - def getLemma(self, **args): - """get lemma""" - return self.template.fulltextclient.getLemma(**args) - - def getToc(self, **args): - """get toc""" - return self.template.fulltextclient.getToc(**args) - - def getTocPage(self, **args): - """get tocpage""" - return self.template.fulltextclient.getTocPage(**args) - security.declareProtected('View','thumbs_rss') def thumbs_rss(self,mode,url,viewMode="auto",start=None,pn=1): ''' @@ -217,7 +162,7 @@ class documentViewer(Folder): pt = getattr(self.template, 'thumbs_main_rss') if viewMode=="auto": # automodus gewaehlt - if docinfo.has_key("textURL") or docinfo.has_key('textURLPath'): #texturl gesetzt und textViewer konfiguriert + if docinfo.get("textURL",'') and self.textViewerUrl: #texturl gesetzt und textViewer konfiguriert viewMode="text" else: viewMode="images" @@ -243,16 +188,17 @@ class documentViewer(Folder): return "ERROR: template folder missing!" if not getattr(self, 'digilibBaseUrl', None): - self.digilibBaseUrl = self.findDigilibUrl() or "http://digilib.mpiwg-berlin.mpg.de/digitallibrary" + self.digilibBaseUrl = self.findDigilibUrl() or "http://nausikaa.mpiwg-berlin.mpg.de/digitallibrary" docinfo = self.getDocinfo(mode=mode,url=url) + if tocMode != "thumbs": # get table of contents docinfo = self.getToc(mode=tocMode, docinfo=docinfo) if viewMode=="auto": # automodus gewaehlt - if docinfo.has_key('textURL') or docinfo.has_key('textURLPath'): #texturl gesetzt und textViewer konfiguriert + if docinfo.get("textURL",''): #texturl gesetzt und textViewer konfiguriert viewMode="text_dict" else: viewMode="images" @@ -266,8 +212,8 @@ class documentViewer(Folder): ret="" if mk is None: return "" - if not isinstance(mk, list): - mk=[mk] + if type(mk) is not ListType: + mk=[mk] for m in mk: ret+="mk=%s"%m return ret @@ -305,8 +251,7 @@ class documentViewer(Folder): params["url"] = getParentDir(params["url"]) # quote values and assemble into query string - #ps = "&".join(["%s=%s"%(k,urllib.quote(v)) for (k, v) in params.items()]) - ps = urllib.urlencode(params) + ps = "&".join(["%s=%s"%(k,urllib.quote(v)) for (k, v) in params.items()]) url=self.REQUEST['URL1']+"?"+ps return url @@ -359,6 +304,7 @@ class documentViewer(Folder): def getDirinfoFromDigilib(self,path,docinfo=None,cut=0): """gibt param von dlInfo aus""" + num_retries = 3 if docinfo is None: docinfo = {} @@ -370,11 +316,17 @@ class documentViewer(Folder): logging.debug("documentViewer (getparamfromdigilib) dirInfo from %s"%(infoUrl)) - txt = getHttpData(infoUrl) - if txt is None: + for cnt in range(num_retries): + try: + # dom = NonvalidatingReader.parseUri(imageUrl) + txt=urllib.urlopen(infoUrl).read() + dom = Parse(txt) + break + except: + logging.error("documentViewer (getdirinfofromdigilib) error reading %s (try %d)"%(infoUrl,cnt)) + else: raise IOError("Unable to get dir-info from %s"%(infoUrl)) - - dom = Parse(txt) + sizes=dom.xpath("//dir/size") logging.debug("documentViewer (getparamfromdigilib) dirInfo:size"%sizes) @@ -390,6 +342,7 @@ class documentViewer(Folder): def getIndexMeta(self, url): """returns dom of index.meta document at url""" + num_retries = 3 dom = None metaUrl = None if url.startswith("http://"): @@ -401,17 +354,25 @@ class documentViewer(Folder): metaUrl=server+url.replace("/mpiwg/online","") if not metaUrl.endswith("index.meta"): metaUrl += "/index.meta" + logging.debug("METAURL: %s"%metaUrl) + for cnt in range(num_retries): + try: + # patch dirk encoding fehler treten dann nicht mehr auf + # dom = NonvalidatingReader.parseUri(metaUrl) + txt=urllib.urlopen(metaUrl).read() + dom = Parse(txt) + break + except: + logging.error("ERROR documentViewer (getIndexMeta) %s (%s)"%sys.exc_info()[0:2]) - logging.debug("(getIndexMeta): METAURL: %s"%metaUrl) - txt=getHttpData(metaUrl) - if txt is None: + if dom is None: raise IOError("Unable to read index meta from %s"%(url)) - - dom = Parse(txt) + return dom def getPresentationInfoXML(self, url): """returns dom of info.xml document at url""" + num_retries = 3 dom = None metaUrl = None if url.startswith("http://"): @@ -422,11 +383,19 @@ class documentViewer(Folder): server=self.digilibBaseUrl+"/servlet/Texter?fn=" metaUrl=server+url.replace("/mpiwg/online","") - txt=getHttpData(metaUrl) - if txt is None: + for cnt in range(num_retries): + try: + # patch dirk encoding fehler treten dann nicht mehr auf + # dom = NonvalidatingReader.parseUri(metaUrl) + txt=urllib.urlopen(metaUrl).read() + dom = Parse(txt) + break + except: + logging.error("ERROR documentViewer (getPresentationInfoXML) %s (%s)"%sys.exc_info()[0:2]) + + if dom is None: raise IOError("Unable to read infoXMLfrom %s"%(url)) - - dom = Parse(txt) + return dom @@ -568,8 +537,7 @@ class documentViewer(Folder): if viewerUrls and (len(viewerUrls) > 0): viewerUrl = getTextFromNode(viewerUrls[0]) docinfo['viewerURL'] = viewerUrl - - # old style text URL + textUrls = dom.xpath("//texttool/text") if textUrls and (len(textUrls) > 0): textUrl = getTextFromNode(textUrls[0]) @@ -581,14 +549,13 @@ class documentViewer(Folder): docinfo['textURL'] = textUrl - # new style text-url-path textUrls = dom.xpath("//texttool/text-url-path") if textUrls and (len(textUrls) > 0): textUrl = getTextFromNode(textUrls[0]) docinfo['textURLPath'] = textUrl if not docinfo['imagePath']: # text-only, no page images - docinfo = self.getNumPages(docinfo) + docinfo = self.getNumPages(docinfo) #im moment einfach auf eins setzen, navigation ueber die thumbs geht natuerlich nicht presentationUrls = dom.xpath("//texttool/presentation") docinfo = self.getBibinfoFromIndexMeta(url, docinfo=docinfo, dom=dom) # get info von bib tag