rename logger fields to logger parts

git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@2579 e7d03fd6-7b0d-0410-9947-9c21f3af8025
This commit is contained in:
calvin 2005-05-08 20:06:40 +00:00
parent afa50a0c3a
commit 6ff56a53cc
8 changed files with 109 additions and 109 deletions

View file

@ -51,15 +51,15 @@ class Logger (object):
def __init__ (self, **args):
"""
Initialize a logger, looking for field restrictions in kwargs.
Initialize a logger, looking for part restrictions in kwargs.
"""
# what log fields should be in output
self.logfields = None # log all fields
if args.has_key('fields'):
if "all" not in args['fields']:
# only log given fields
self.logfields = args['fields']
# number of spaces before log fields for alignment
# what log parts should be in output
self.logparts = None # log all parts
if args.has_key('parts'):
if "all" not in args['parts']:
# only log given parts
self.logparts = args['parts']
# number of spaces before log parts for alignment
self.logspaces = {}
# maximum indent of spaces for alignment
self.max_indent = 0
@ -154,24 +154,24 @@ class Logger (object):
self.write(s)
self.write(unicode(os.linesep), **args)
def has_field (self, name):
def has_part (self, name):
"""
See if given field name will be logged.
See if given part name will be logged.
"""
if self.logfields is None:
# log all fields
if self.logparts is None:
# log all parts
return True
return name in self.logfields
return name in self.logparts
def field (self, name):
def part (self, name):
"""
Return translated field name.
Return translated part name.
"""
return _(Fields[name])
def spaces (self, name):
"""
Return indent of spaces for given field name.
Return indent of spaces for given part name.
"""
return self.logspaces[name]
@ -179,16 +179,16 @@ class Logger (object):
"""
Start log output.
"""
# map with spaces between field name and value
if self.logfields is None:
fields = Fields.keys()
# map with spaces between part name and value
if self.logparts is None:
parts = Fields.keys()
else:
fields = self.logfields
values = [self.field(x) for x in fields]
# maximum indent for localized log field names
parts = self.logparts
values = [self.part(x) for x in parts]
# maximum indent for localized log part names
self.max_indent = max([len(x) for x in values])+1
for key in fields:
numspaces = (self.max_indent - len(self.field(key)))
for key in parts:
numspaces = (self.max_indent - len(self.part(key)))
self.logspaces[key] = u" " * numspaces
def log_filter_url (self, url_data, do_print):

View file

@ -57,7 +57,7 @@ class CSVLogger (linkcheck.logger.Logger):
return
self.starttime = time.time()
row = []
if self.has_field("intro"):
if self.has_part("intro"):
self.comment(_("created by %s at %s") % \
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
@ -123,7 +123,7 @@ class CSVLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.stoptime = time.time()
if self.has_field("outro"):
if self.has_part("outro"):
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)") % \
(linkcheck.strformat.strtime(self.stoptime),

View file

@ -46,7 +46,7 @@ class DOTLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.starttime = time.time()
if self.has_field("intro"):
if self.has_part("intro"):
self.comment(_("created by %s at %s") % \
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
@ -78,22 +78,22 @@ class DOTLogger (linkcheck.logger.Logger):
self.nodes[node.url] = node
self.nodeid += 1
self.writeln(u" %d [" % node.id)
if self.has_field("realurl"):
if self.has_part("realurl"):
self.writeln(u' href="%s",' % dotquote(node.url))
if node.dltime >= 0 and self.has_field("dltime"):
if node.dltime >= 0 and self.has_part("dltime"):
self.writeln(u" dltime=%d," % node.dltime)
if node.dlsize >= 0 and self.has_field("dlsize"):
if node.dlsize >= 0 and self.has_part("dlsize"):
self.writeln(u" dlsize=%d," % node.dlsize)
if node.checktime and self.has_field("checktime"):
if node.checktime and self.has_part("checktime"):
self.writeln(u" checktime=%d," % node.checktime)
if self.has_field("extern"):
if self.has_part("extern"):
self.writeln(u" extern=%d," % (node.extern and 1 or 0))
self.writeln(u" ];")
def write_edges (self):
"""
Write all edges we can find in the graph in a brute-force
manner. Better would be a mapping of parent urls.
manner. Better would be a mapping of parent URLs.
"""
for node in self.nodes.values():
if self.nodes.has_key(node.parent_url):
@ -101,7 +101,7 @@ class DOTLogger (linkcheck.logger.Logger):
target = node.id
self.writeln(u" %d -> %d [" % (source, target))
self.writeln(u' label="%s",' % dotedge(node.name))
if self.has_field("result"):
if self.has_part("result"):
self.writeln(u" valid=%d," % (node.valid and 1 or 0))
self.writeln(u" ];")
self.flush()
@ -114,7 +114,7 @@ class DOTLogger (linkcheck.logger.Logger):
return
self.write_edges()
self.writeln(u"}")
if self.has_field("outro"):
if self.has_part("outro"):
self.stoptime = time.time()
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)")%\

View file

@ -46,7 +46,7 @@ class GMLLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.starttime = time.time()
if self.has_field("intro"):
if self.has_part("intro"):
self.comment(_("created by %s at %s") % \
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
@ -80,32 +80,32 @@ class GMLLogger (linkcheck.logger.Logger):
self.nodeid += 1
self.writeln(u" node [")
self.writeln(u" id %d" % node.id)
if self.has_field("realurl"):
if self.has_part("realurl"):
self.writeln(u' label "%s"' % node.url)
if node.dltime >= 0 and self.has_field("dltime"):
if node.dltime >= 0 and self.has_part("dltime"):
self.writeln(u" dltime %d" % node.dltime)
if node.dlsize >= 0 and self.has_field("dlsize"):
if node.dlsize >= 0 and self.has_part("dlsize"):
self.writeln(u" dlsize %d" % node.dlsize)
if node.checktime and self.has_field("checktime"):
if node.checktime and self.has_part("checktime"):
self.writeln(u" checktime %d" % node.checktime)
if self.has_field("extern"):
if self.has_part("extern"):
self.writeln(u" extern %d" % (node.extern and 1 or 0))
self.writeln(u" ]")
def write_edges (self):
"""
Write all edges we can find in the graph in a brute-force
manner. Better would be a mapping of parent urls.
manner. Better would be a mapping of parent URLs.
"""
for node in self.nodes.values():
if self.nodes.has_key(node.parent_url):
self.writeln(u" edge [")
self.writeln(u' label "%s"' % node.base_url)
if self.has_field("parenturl"):
if self.has_part("parenturl"):
self.writeln(u" source %d" % \
self.nodes[node.parent_url].id)
self.writeln(u" target %d" % node.id)
if self.has_field("result"):
if self.has_part("result"):
self.writeln(u" valid %d" % (node.valid and 1 or 0))
self.writeln(u" ]")
self.flush()
@ -118,7 +118,7 @@ class GMLLogger (linkcheck.logger.Logger):
return
self.write_edges()
self.writeln(u"]")
if self.has_field("outro"):
if self.has_part("outro"):
self.stoptime = time.time()
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)")%\

View file

@ -72,11 +72,11 @@ class HtmlLogger (linkcheck.logger.Logger):
self.colorerror = args['colorerror']
self.colorok = args['colorok']
def field (self, name):
def part (self, name):
"""
Return non-space-breakable field name.
Return non-space-breakable part name.
"""
return super(HtmlLogger, self).field(name).replace(" ", " ")
return super(HtmlLogger, self).part(name).replace(" ", " ")
def comment (self, s, **args):
"""
@ -97,7 +97,7 @@ class HtmlLogger (linkcheck.logger.Logger):
self.write(HTML_HEADER % (linkcheck.configuration.App,
self.colorbackground,
self.colorlink, self.colorlink, self.colorlink))
if self.has_field('intro'):
if self.has_part('intro'):
self.write(u"<center><h2>"+linkcheck.configuration.App+
"</h2></center><br><blockquote>"+
linkcheck.configuration.Freeware+"<br><br>"+
@ -114,27 +114,27 @@ class HtmlLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.write_table_start()
if self.has_field("url"):
if self.has_part("url"):
self.write_url(url_data)
if url_data.name and self.has_field("name"):
if url_data.name and self.has_part("name"):
self.write_name(url_data)
if url_data.parent_url and self.has_field("parenturl"):
if url_data.parent_url and self.has_part("parenturl"):
self.write_parent(url_data)
if url_data.base_ref and self.has_field("base"):
if url_data.base_ref and self.has_part("base"):
self.write_base(url_data)
if url_data.url and self.has_field("realurl"):
if url_data.url and self.has_part("realurl"):
self.write_real(url_data)
if url_data.dltime >= 0 and self.has_field("dltime"):
if url_data.dltime >= 0 and self.has_part("dltime"):
self.write_dltime(url_data)
if url_data.dlsize >= 0 and self.has_field("dlsize"):
if url_data.dlsize >= 0 and self.has_part("dlsize"):
self.write_dlsize(url_data)
if url_data.checktime and self.has_field("checktime"):
if url_data.checktime and self.has_part("checktime"):
self.write_checktime(url_data)
if url_data.info and self.has_field("info"):
if url_data.info and self.has_part("info"):
self.write_info(url_data)
if url_data.warning and self.has_field("warning"):
if url_data.warning and self.has_part("warning"):
self.write_warning(url_data)
if self.has_field("result"):
if self.has_part("result"):
self.write_result(url_data)
self.write_table_end()
self.flush()
@ -167,7 +167,7 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
self.writeln(u"<tr>")
self.writeln(u"<td bgcolor=\""+self.colorurl+u"\">"+
self.field("url")+u"</td>")
self.part("url")+u"</td>")
self.write(u"<td bgcolor=\""+self.colorurl+u"\">"+
cgi.escape(repr(url_data.base_url)[1:]))
if url_data.cached:
@ -178,14 +178,14 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
Write url_data.name.
"""
self.writeln(u"<tr><td>"+self.field("name")+u"</td><td>"+
self.writeln(u"<tr><td>"+self.part("name")+u"</td><td>"+
cgi.escape(repr(url_data.name)[1:])+u"</td></tr>")
def write_parent (self, url_data):
"""
Write url_data.parent_url.
"""
self.write(u"<tr><td>"+self.field("parenturl")+
self.write(u"<tr><td>"+self.part("parenturl")+
u'</td><td><a target="top" href="'+
url_data.parent_url+u'">'+
cgi.escape(url_data.parent_url)+u"</a>")
@ -204,14 +204,14 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
Write url_data.base_ref.
"""
self.writeln(u"<tr><td>"+self.field("base")+u"</td><td>"+
self.writeln(u"<tr><td>"+self.part("base")+u"</td><td>"+
cgi.escape(url_data.base_ref)+u"</td></tr>")
def write_real (self, url_data):
"""
Write url_data.url.
"""
self.writeln("<tr><td>"+self.field("realurl")+u"</td><td>"+
self.writeln("<tr><td>"+self.part("realurl")+u"</td><td>"+
u'<a target="top" href="'+url_data.url+
u'">'+cgi.escape(url_data.url)+u"</a></td></tr>")
@ -219,7 +219,7 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
Write url_data.dltime.
"""
self.writeln(u"<tr><td>"+self.field("dltime")+u"</td><td>"+
self.writeln(u"<tr><td>"+self.part("dltime")+u"</td><td>"+
(_("%.3f seconds") % url_data.dltime)+
u"</td></tr>")
@ -227,7 +227,7 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
Write url_data.dlsize.
"""
self.writeln(u"<tr><td>"+self.field("dlsize")+u"</td><td>"+
self.writeln(u"<tr><td>"+self.part("dlsize")+u"</td><td>"+
linkcheck.strformat.strsize(url_data.dlsize)+
u"</td></tr>")
@ -235,7 +235,7 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
Write url_data.checktime.
"""
self.writeln(u"<tr><td>"+self.field("checktime")+u"</td><td>"+
self.writeln(u"<tr><td>"+self.part("checktime")+u"</td><td>"+
(_("%.3f seconds") % url_data.checktime)+u"</td></tr>")
def write_info (self, url_data):
@ -243,7 +243,7 @@ class HtmlLogger (linkcheck.logger.Logger):
Write url_data.info.
"""
text = os.linesep.join(url_data.info)
self.writeln(u"<tr><td valign=\"top\">"+self.field("info")+
self.writeln(u"<tr><td valign=\"top\">"+self.part("info")+
u"</td><td>"+cgi.escape(text).replace(os.linesep, "<br>")+
u"</td></tr>")
@ -254,7 +254,7 @@ class HtmlLogger (linkcheck.logger.Logger):
sep = u"<br>"+os.linesep
text = sep.join([cgi.escape(x) for x in url_data.warning])
self.writeln(u"<tr><td bgcolor=\""+self.colorwarning+u"\" "+
u"valign=\"top\">"+self.field("warning")+
u"valign=\"top\">"+self.part("warning")+
u"</td><td bgcolor=\""+self.colorwarning+u"\">"+
text+u"</td></tr>")
@ -264,11 +264,11 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
if url_data.valid:
self.write(u"<tr><td bgcolor=\""+self.colorok+u"\">"+
self.field("result")+u"</td><td bgcolor=\""+self.colorok+u"\">")
self.part("result")+u"</td><td bgcolor=\""+self.colorok+u"\">")
self.write(_("Valid"))
else:
self.write(u"<tr><td bgcolor=\""+self.colorerror+u"\">"+
self.field("result")+u"</td><td bgcolor=\""+self.colorerror+u"\">")
self.part("result")+u"</td><td bgcolor=\""+self.colorerror+u"\">")
self.write(_("Error"))
if url_data.result:
self.write(u": "+cgi.escape(url_data.result))
@ -280,7 +280,7 @@ class HtmlLogger (linkcheck.logger.Logger):
"""
if self.fd is None:
return
if self.has_field("outro"):
if self.has_part("outro"):
self.writeln()
self.write(_("That's it.")+" ")
if self.number >= 0:

View file

@ -77,7 +77,7 @@ class SQLLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.starttime = time.time()
if self.has_field("intro"):
if self.has_part("intro"):
self.comment(_("created by %s at %s") % \
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
@ -142,7 +142,7 @@ class SQLLogger (linkcheck.logger.Logger):
"""
if self.fd is None:
return
if self.has_field("outro"):
if self.has_part("outro"):
self.stoptime = time.time()
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)") % \

View file

@ -92,7 +92,7 @@ class TextLogger (linkcheck.logger.Logger):
if self.fd is None:
return
self.starttime = time.time()
if self.has_field('intro'):
if self.has_part('intro'):
self.writeln(linkcheck.configuration.AppInfo)
self.writeln(linkcheck.configuration.Freeware)
self.writeln(_("Get the newest version at %s") %
@ -111,27 +111,27 @@ class TextLogger (linkcheck.logger.Logger):
"""
if self.fd is None:
return
if self.has_field('url'):
if self.has_part('url'):
self.write_url(url_data)
if url_data.name and self.has_field('name'):
if url_data.name and self.has_part('name'):
self.write_name(url_data)
if url_data.parent_url and self.has_field('parenturl'):
if url_data.parent_url and self.has_part('parenturl'):
self.write_parent(url_data)
if url_data.base_ref and self.has_field('base'):
if url_data.base_ref and self.has_part('base'):
self.write_base(url_data)
if url_data.url and self.has_field('realurl'):
if url_data.url and self.has_part('realurl'):
self.write_real(url_data)
if url_data.dltime >= 0 and self.has_field('dltime'):
if url_data.dltime >= 0 and self.has_part('dltime'):
self.write_dltime(url_data)
if url_data.dlsize >= 0 and self.has_field('dlsize'):
if url_data.dlsize >= 0 and self.has_part('dlsize'):
self.write_dlsize(url_data)
if url_data.checktime and self.has_field('checktime'):
if url_data.checktime and self.has_part('checktime'):
self.write_checktime(url_data)
if url_data.info and self.has_field('info'):
if url_data.info and self.has_part('info'):
self.write_info(url_data)
if url_data.warning and self.has_field('warning'):
if url_data.warning and self.has_part('warning'):
self.write_warning(url_data)
if self.has_field('result'):
if self.has_part('result'):
self.write_result(url_data)
self.flush()
@ -140,7 +140,7 @@ class TextLogger (linkcheck.logger.Logger):
Write url_data.base_url.
"""
self.writeln()
self.write(self.field('url') + self.spaces('url'))
self.write(self.part('url') + self.spaces('url'))
txt = unicode(repr(url_data.base_url)[1:])
if url_data.cached:
txt += _(" (cached)")
@ -150,14 +150,14 @@ class TextLogger (linkcheck.logger.Logger):
"""
Write url_data.name.
"""
self.write(self.field("name") + self.spaces("name"))
self.write(self.part("name") + self.spaces("name"))
self.writeln(unicode(repr(url_data.name)[1:]), color=self.colorname)
def write_parent (self, url_data):
"""
Write url_data.parent_url.
"""
self.write(self.field('parenturl') + self.spaces("parenturl"))
self.write(self.part('parenturl') + self.spaces("parenturl"))
txt = url_data.parent_url
txt += _(", line %d") % url_data.line
txt += _(", col %d") % url_data.column
@ -167,21 +167,21 @@ class TextLogger (linkcheck.logger.Logger):
"""
Write url_data.base_ref.
"""
self.write(self.field("base") + self.spaces("base"))
self.write(self.part("base") + self.spaces("base"))
self.writeln(url_data.base_ref, color=self.colorbase)
def write_real (self, url_data):
"""
Write url_data.url.
"""
self.write(self.field("realurl") + self.spaces("realurl"))
self.write(self.part("realurl") + self.spaces("realurl"))
self.writeln(unicode(url_data.url), color=self.colorreal)
def write_dltime (self, url_data):
"""
Write url_data.dltime.
"""
self.write(self.field("dltime") + self.spaces("dltime"))
self.write(self.part("dltime") + self.spaces("dltime"))
self.writeln(_("%.3f seconds") % url_data.dltime,
color=self.colordltime)
@ -189,7 +189,7 @@ class TextLogger (linkcheck.logger.Logger):
"""
Write url_data.dlsize.
"""
self.write(self.field("dlsize") + self.spaces("dlsize"))
self.write(self.part("dlsize") + self.spaces("dlsize"))
self.writeln(linkcheck.strformat.strsize(url_data.dlsize),
color=self.colordlsize)
@ -197,7 +197,7 @@ class TextLogger (linkcheck.logger.Logger):
"""
Write url_data.checktime.
"""
self.write(self.field("checktime") + self.spaces("checktime"))
self.write(self.part("checktime") + self.spaces("checktime"))
self.writeln(_("%.3f seconds") % url_data.checktime,
color=self.colordltime)
@ -205,21 +205,21 @@ class TextLogger (linkcheck.logger.Logger):
"""
Write url_data.info.
"""
self.write(self.field("info") + self.spaces("info"))
self.write(self.part("info") + self.spaces("info"))
self.writeln(self.wrap(url_data.info, 65), color=self.colorinfo)
def write_warning (self, url_data):
"""
Write url_data.warning.
"""
self.write(self.field("warning") + self.spaces("warning"))
self.write(self.part("warning") + self.spaces("warning"))
self.writeln(self.wrap(url_data.warning, 65), color=self.colorwarning)
def write_result (self, url_data):
"""
Write url_data.result.
"""
self.write(self.field("result") + self.spaces("result"))
self.write(self.part("result") + self.spaces("result"))
if url_data.valid:
color = self.colorvalid
self.write(_("Valid"), color=color)
@ -236,7 +236,7 @@ class TextLogger (linkcheck.logger.Logger):
"""
if self.fd is None:
return
if self.has_field('outro'):
if self.has_part('outro'):
self.writeln()
self.write(_("That's it.")+" ")
if self.number >= 0:

View file

@ -93,7 +93,7 @@ class XMLLogger (linkcheck.logger.Logger):
return
self.starttime = time.time()
self.writeln(u'<?xml version="1.0"?>')
if self.has_field("intro"):
if self.has_part("intro"):
self.comment(_("created by %s at %s") %
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
@ -119,17 +119,17 @@ class XMLLogger (linkcheck.logger.Logger):
self.nodes[node.url] = node
self.nodeid += 1
self.writeln(u' <node name="%d">' % node.id)
if self.has_field("realurl"):
if self.has_part("realurl"):
self.writeln(u" <label>%s</label>" % xmlquote(node.url))
self.writeln(u" <data>")
if node.dltime >= 0 and self.has_field("dltime"):
if node.dltime >= 0 and self.has_part("dltime"):
self.writeln(u" <dltime>%f</dltime>" % node.dltime)
if node.dlsize >= 0 and self.has_field("dlsize"):
if node.dlsize >= 0 and self.has_part("dlsize"):
self.writeln(u" <dlsize>%d</dlsize>" % node.dlsize)
if node.checktime and self.has_field("checktime"):
if node.checktime and self.has_part("checktime"):
self.writeln(u" <checktime>%f</checktime>" %
node.checktime)
if self.has_field("extern"):
if self.has_part("extern"):
self.writeln(u" <extern>%d</extern>" %
(node.extern and 1 or 0))
self.writeln(u" </data>")
@ -139,18 +139,18 @@ class XMLLogger (linkcheck.logger.Logger):
def write_edges (self):
"""
Write all edges we can find in the graph in a brute-force
manner. Better would be a mapping of parent urls.
manner. Better would be a mapping of parent URLs.
"""
for node in self.nodes.values():
if self.nodes.has_key(node.parent_url):
self.write(u" <edge")
self.write(u' source="%d"' % self.nodes[node.parent_url].id)
self.writeln(u' target="%d">' % node.id)
if self.has_field("url"):
if self.has_part("url"):
self.writeln(u" <label>%s</label>" % \
xmlquote(node.base_url))
self.writeln(u" <data>")
if self.has_field("result"):
if self.has_part("result"):
self.writeln(u" <valid>%d</valid>" % \
(node.valid and 1 or 0))
self.writeln(u" </data>")
@ -165,7 +165,7 @@ class XMLLogger (linkcheck.logger.Logger):
return
self.writeln(u"</graph>")
self.writeln(u"</GraphXML>")
if self.has_field("outro"):
if self.has_part("outro"):
self.stoptime = time.time()
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)") % \