mirror of
https://github.com/Hopiu/linkchecker.git
synced 2026-03-23 01:10:27 +00:00
rename logger fields to logger parts
git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@2579 e7d03fd6-7b0d-0410-9947-9c21f3af8025
This commit is contained in:
parent
afa50a0c3a
commit
6ff56a53cc
8 changed files with 109 additions and 109 deletions
|
|
@ -51,15 +51,15 @@ class Logger (object):
|
|||
|
||||
def __init__ (self, **args):
|
||||
"""
|
||||
Initialize a logger, looking for field restrictions in kwargs.
|
||||
Initialize a logger, looking for part restrictions in kwargs.
|
||||
"""
|
||||
# what log fields should be in output
|
||||
self.logfields = None # log all fields
|
||||
if args.has_key('fields'):
|
||||
if "all" not in args['fields']:
|
||||
# only log given fields
|
||||
self.logfields = args['fields']
|
||||
# number of spaces before log fields for alignment
|
||||
# what log parts should be in output
|
||||
self.logparts = None # log all parts
|
||||
if args.has_key('parts'):
|
||||
if "all" not in args['parts']:
|
||||
# only log given parts
|
||||
self.logparts = args['parts']
|
||||
# number of spaces before log parts for alignment
|
||||
self.logspaces = {}
|
||||
# maximum indent of spaces for alignment
|
||||
self.max_indent = 0
|
||||
|
|
@ -154,24 +154,24 @@ class Logger (object):
|
|||
self.write(s)
|
||||
self.write(unicode(os.linesep), **args)
|
||||
|
||||
def has_field (self, name):
|
||||
def has_part (self, name):
|
||||
"""
|
||||
See if given field name will be logged.
|
||||
See if given part name will be logged.
|
||||
"""
|
||||
if self.logfields is None:
|
||||
# log all fields
|
||||
if self.logparts is None:
|
||||
# log all parts
|
||||
return True
|
||||
return name in self.logfields
|
||||
return name in self.logparts
|
||||
|
||||
def field (self, name):
|
||||
def part (self, name):
|
||||
"""
|
||||
Return translated field name.
|
||||
Return translated part name.
|
||||
"""
|
||||
return _(Fields[name])
|
||||
|
||||
def spaces (self, name):
|
||||
"""
|
||||
Return indent of spaces for given field name.
|
||||
Return indent of spaces for given part name.
|
||||
"""
|
||||
return self.logspaces[name]
|
||||
|
||||
|
|
@ -179,16 +179,16 @@ class Logger (object):
|
|||
"""
|
||||
Start log output.
|
||||
"""
|
||||
# map with spaces between field name and value
|
||||
if self.logfields is None:
|
||||
fields = Fields.keys()
|
||||
# map with spaces between part name and value
|
||||
if self.logparts is None:
|
||||
parts = Fields.keys()
|
||||
else:
|
||||
fields = self.logfields
|
||||
values = [self.field(x) for x in fields]
|
||||
# maximum indent for localized log field names
|
||||
parts = self.logparts
|
||||
values = [self.part(x) for x in parts]
|
||||
# maximum indent for localized log part names
|
||||
self.max_indent = max([len(x) for x in values])+1
|
||||
for key in fields:
|
||||
numspaces = (self.max_indent - len(self.field(key)))
|
||||
for key in parts:
|
||||
numspaces = (self.max_indent - len(self.part(key)))
|
||||
self.logspaces[key] = u" " * numspaces
|
||||
|
||||
def log_filter_url (self, url_data, do_print):
|
||||
|
|
|
|||
|
|
@ -57,7 +57,7 @@ class CSVLogger (linkcheck.logger.Logger):
|
|||
return
|
||||
self.starttime = time.time()
|
||||
row = []
|
||||
if self.has_field("intro"):
|
||||
if self.has_part("intro"):
|
||||
self.comment(_("created by %s at %s") % \
|
||||
(linkcheck.configuration.AppName,
|
||||
linkcheck.strformat.strtime(self.starttime)))
|
||||
|
|
@ -123,7 +123,7 @@ class CSVLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.stoptime = time.time()
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
duration = self.stoptime - self.starttime
|
||||
self.comment(_("Stopped checking at %s (%s)") % \
|
||||
(linkcheck.strformat.strtime(self.stoptime),
|
||||
|
|
|
|||
|
|
@ -46,7 +46,7 @@ class DOTLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.starttime = time.time()
|
||||
if self.has_field("intro"):
|
||||
if self.has_part("intro"):
|
||||
self.comment(_("created by %s at %s") % \
|
||||
(linkcheck.configuration.AppName,
|
||||
linkcheck.strformat.strtime(self.starttime)))
|
||||
|
|
@ -78,22 +78,22 @@ class DOTLogger (linkcheck.logger.Logger):
|
|||
self.nodes[node.url] = node
|
||||
self.nodeid += 1
|
||||
self.writeln(u" %d [" % node.id)
|
||||
if self.has_field("realurl"):
|
||||
if self.has_part("realurl"):
|
||||
self.writeln(u' href="%s",' % dotquote(node.url))
|
||||
if node.dltime >= 0 and self.has_field("dltime"):
|
||||
if node.dltime >= 0 and self.has_part("dltime"):
|
||||
self.writeln(u" dltime=%d," % node.dltime)
|
||||
if node.dlsize >= 0 and self.has_field("dlsize"):
|
||||
if node.dlsize >= 0 and self.has_part("dlsize"):
|
||||
self.writeln(u" dlsize=%d," % node.dlsize)
|
||||
if node.checktime and self.has_field("checktime"):
|
||||
if node.checktime and self.has_part("checktime"):
|
||||
self.writeln(u" checktime=%d," % node.checktime)
|
||||
if self.has_field("extern"):
|
||||
if self.has_part("extern"):
|
||||
self.writeln(u" extern=%d," % (node.extern and 1 or 0))
|
||||
self.writeln(u" ];")
|
||||
|
||||
def write_edges (self):
|
||||
"""
|
||||
Write all edges we can find in the graph in a brute-force
|
||||
manner. Better would be a mapping of parent urls.
|
||||
manner. Better would be a mapping of parent URLs.
|
||||
"""
|
||||
for node in self.nodes.values():
|
||||
if self.nodes.has_key(node.parent_url):
|
||||
|
|
@ -101,7 +101,7 @@ class DOTLogger (linkcheck.logger.Logger):
|
|||
target = node.id
|
||||
self.writeln(u" %d -> %d [" % (source, target))
|
||||
self.writeln(u' label="%s",' % dotedge(node.name))
|
||||
if self.has_field("result"):
|
||||
if self.has_part("result"):
|
||||
self.writeln(u" valid=%d," % (node.valid and 1 or 0))
|
||||
self.writeln(u" ];")
|
||||
self.flush()
|
||||
|
|
@ -114,7 +114,7 @@ class DOTLogger (linkcheck.logger.Logger):
|
|||
return
|
||||
self.write_edges()
|
||||
self.writeln(u"}")
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
self.stoptime = time.time()
|
||||
duration = self.stoptime - self.starttime
|
||||
self.comment(_("Stopped checking at %s (%s)")%\
|
||||
|
|
|
|||
|
|
@ -46,7 +46,7 @@ class GMLLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.starttime = time.time()
|
||||
if self.has_field("intro"):
|
||||
if self.has_part("intro"):
|
||||
self.comment(_("created by %s at %s") % \
|
||||
(linkcheck.configuration.AppName,
|
||||
linkcheck.strformat.strtime(self.starttime)))
|
||||
|
|
@ -80,32 +80,32 @@ class GMLLogger (linkcheck.logger.Logger):
|
|||
self.nodeid += 1
|
||||
self.writeln(u" node [")
|
||||
self.writeln(u" id %d" % node.id)
|
||||
if self.has_field("realurl"):
|
||||
if self.has_part("realurl"):
|
||||
self.writeln(u' label "%s"' % node.url)
|
||||
if node.dltime >= 0 and self.has_field("dltime"):
|
||||
if node.dltime >= 0 and self.has_part("dltime"):
|
||||
self.writeln(u" dltime %d" % node.dltime)
|
||||
if node.dlsize >= 0 and self.has_field("dlsize"):
|
||||
if node.dlsize >= 0 and self.has_part("dlsize"):
|
||||
self.writeln(u" dlsize %d" % node.dlsize)
|
||||
if node.checktime and self.has_field("checktime"):
|
||||
if node.checktime and self.has_part("checktime"):
|
||||
self.writeln(u" checktime %d" % node.checktime)
|
||||
if self.has_field("extern"):
|
||||
if self.has_part("extern"):
|
||||
self.writeln(u" extern %d" % (node.extern and 1 or 0))
|
||||
self.writeln(u" ]")
|
||||
|
||||
def write_edges (self):
|
||||
"""
|
||||
Write all edges we can find in the graph in a brute-force
|
||||
manner. Better would be a mapping of parent urls.
|
||||
manner. Better would be a mapping of parent URLs.
|
||||
"""
|
||||
for node in self.nodes.values():
|
||||
if self.nodes.has_key(node.parent_url):
|
||||
self.writeln(u" edge [")
|
||||
self.writeln(u' label "%s"' % node.base_url)
|
||||
if self.has_field("parenturl"):
|
||||
if self.has_part("parenturl"):
|
||||
self.writeln(u" source %d" % \
|
||||
self.nodes[node.parent_url].id)
|
||||
self.writeln(u" target %d" % node.id)
|
||||
if self.has_field("result"):
|
||||
if self.has_part("result"):
|
||||
self.writeln(u" valid %d" % (node.valid and 1 or 0))
|
||||
self.writeln(u" ]")
|
||||
self.flush()
|
||||
|
|
@ -118,7 +118,7 @@ class GMLLogger (linkcheck.logger.Logger):
|
|||
return
|
||||
self.write_edges()
|
||||
self.writeln(u"]")
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
self.stoptime = time.time()
|
||||
duration = self.stoptime - self.starttime
|
||||
self.comment(_("Stopped checking at %s (%s)")%\
|
||||
|
|
|
|||
|
|
@ -72,11 +72,11 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
self.colorerror = args['colorerror']
|
||||
self.colorok = args['colorok']
|
||||
|
||||
def field (self, name):
|
||||
def part (self, name):
|
||||
"""
|
||||
Return non-space-breakable field name.
|
||||
Return non-space-breakable part name.
|
||||
"""
|
||||
return super(HtmlLogger, self).field(name).replace(" ", " ")
|
||||
return super(HtmlLogger, self).part(name).replace(" ", " ")
|
||||
|
||||
def comment (self, s, **args):
|
||||
"""
|
||||
|
|
@ -97,7 +97,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
self.write(HTML_HEADER % (linkcheck.configuration.App,
|
||||
self.colorbackground,
|
||||
self.colorlink, self.colorlink, self.colorlink))
|
||||
if self.has_field('intro'):
|
||||
if self.has_part('intro'):
|
||||
self.write(u"<center><h2>"+linkcheck.configuration.App+
|
||||
"</h2></center><br><blockquote>"+
|
||||
linkcheck.configuration.Freeware+"<br><br>"+
|
||||
|
|
@ -114,27 +114,27 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.write_table_start()
|
||||
if self.has_field("url"):
|
||||
if self.has_part("url"):
|
||||
self.write_url(url_data)
|
||||
if url_data.name and self.has_field("name"):
|
||||
if url_data.name and self.has_part("name"):
|
||||
self.write_name(url_data)
|
||||
if url_data.parent_url and self.has_field("parenturl"):
|
||||
if url_data.parent_url and self.has_part("parenturl"):
|
||||
self.write_parent(url_data)
|
||||
if url_data.base_ref and self.has_field("base"):
|
||||
if url_data.base_ref and self.has_part("base"):
|
||||
self.write_base(url_data)
|
||||
if url_data.url and self.has_field("realurl"):
|
||||
if url_data.url and self.has_part("realurl"):
|
||||
self.write_real(url_data)
|
||||
if url_data.dltime >= 0 and self.has_field("dltime"):
|
||||
if url_data.dltime >= 0 and self.has_part("dltime"):
|
||||
self.write_dltime(url_data)
|
||||
if url_data.dlsize >= 0 and self.has_field("dlsize"):
|
||||
if url_data.dlsize >= 0 and self.has_part("dlsize"):
|
||||
self.write_dlsize(url_data)
|
||||
if url_data.checktime and self.has_field("checktime"):
|
||||
if url_data.checktime and self.has_part("checktime"):
|
||||
self.write_checktime(url_data)
|
||||
if url_data.info and self.has_field("info"):
|
||||
if url_data.info and self.has_part("info"):
|
||||
self.write_info(url_data)
|
||||
if url_data.warning and self.has_field("warning"):
|
||||
if url_data.warning and self.has_part("warning"):
|
||||
self.write_warning(url_data)
|
||||
if self.has_field("result"):
|
||||
if self.has_part("result"):
|
||||
self.write_result(url_data)
|
||||
self.write_table_end()
|
||||
self.flush()
|
||||
|
|
@ -167,7 +167,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
self.writeln(u"<tr>")
|
||||
self.writeln(u"<td bgcolor=\""+self.colorurl+u"\">"+
|
||||
self.field("url")+u"</td>")
|
||||
self.part("url")+u"</td>")
|
||||
self.write(u"<td bgcolor=\""+self.colorurl+u"\">"+
|
||||
cgi.escape(repr(url_data.base_url)[1:]))
|
||||
if url_data.cached:
|
||||
|
|
@ -178,14 +178,14 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.name.
|
||||
"""
|
||||
self.writeln(u"<tr><td>"+self.field("name")+u"</td><td>"+
|
||||
self.writeln(u"<tr><td>"+self.part("name")+u"</td><td>"+
|
||||
cgi.escape(repr(url_data.name)[1:])+u"</td></tr>")
|
||||
|
||||
def write_parent (self, url_data):
|
||||
"""
|
||||
Write url_data.parent_url.
|
||||
"""
|
||||
self.write(u"<tr><td>"+self.field("parenturl")+
|
||||
self.write(u"<tr><td>"+self.part("parenturl")+
|
||||
u'</td><td><a target="top" href="'+
|
||||
url_data.parent_url+u'">'+
|
||||
cgi.escape(url_data.parent_url)+u"</a>")
|
||||
|
|
@ -204,14 +204,14 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.base_ref.
|
||||
"""
|
||||
self.writeln(u"<tr><td>"+self.field("base")+u"</td><td>"+
|
||||
self.writeln(u"<tr><td>"+self.part("base")+u"</td><td>"+
|
||||
cgi.escape(url_data.base_ref)+u"</td></tr>")
|
||||
|
||||
def write_real (self, url_data):
|
||||
"""
|
||||
Write url_data.url.
|
||||
"""
|
||||
self.writeln("<tr><td>"+self.field("realurl")+u"</td><td>"+
|
||||
self.writeln("<tr><td>"+self.part("realurl")+u"</td><td>"+
|
||||
u'<a target="top" href="'+url_data.url+
|
||||
u'">'+cgi.escape(url_data.url)+u"</a></td></tr>")
|
||||
|
||||
|
|
@ -219,7 +219,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.dltime.
|
||||
"""
|
||||
self.writeln(u"<tr><td>"+self.field("dltime")+u"</td><td>"+
|
||||
self.writeln(u"<tr><td>"+self.part("dltime")+u"</td><td>"+
|
||||
(_("%.3f seconds") % url_data.dltime)+
|
||||
u"</td></tr>")
|
||||
|
||||
|
|
@ -227,7 +227,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.dlsize.
|
||||
"""
|
||||
self.writeln(u"<tr><td>"+self.field("dlsize")+u"</td><td>"+
|
||||
self.writeln(u"<tr><td>"+self.part("dlsize")+u"</td><td>"+
|
||||
linkcheck.strformat.strsize(url_data.dlsize)+
|
||||
u"</td></tr>")
|
||||
|
||||
|
|
@ -235,7 +235,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.checktime.
|
||||
"""
|
||||
self.writeln(u"<tr><td>"+self.field("checktime")+u"</td><td>"+
|
||||
self.writeln(u"<tr><td>"+self.part("checktime")+u"</td><td>"+
|
||||
(_("%.3f seconds") % url_data.checktime)+u"</td></tr>")
|
||||
|
||||
def write_info (self, url_data):
|
||||
|
|
@ -243,7 +243,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
Write url_data.info.
|
||||
"""
|
||||
text = os.linesep.join(url_data.info)
|
||||
self.writeln(u"<tr><td valign=\"top\">"+self.field("info")+
|
||||
self.writeln(u"<tr><td valign=\"top\">"+self.part("info")+
|
||||
u"</td><td>"+cgi.escape(text).replace(os.linesep, "<br>")+
|
||||
u"</td></tr>")
|
||||
|
||||
|
|
@ -254,7 +254,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
sep = u"<br>"+os.linesep
|
||||
text = sep.join([cgi.escape(x) for x in url_data.warning])
|
||||
self.writeln(u"<tr><td bgcolor=\""+self.colorwarning+u"\" "+
|
||||
u"valign=\"top\">"+self.field("warning")+
|
||||
u"valign=\"top\">"+self.part("warning")+
|
||||
u"</td><td bgcolor=\""+self.colorwarning+u"\">"+
|
||||
text+u"</td></tr>")
|
||||
|
||||
|
|
@ -264,11 +264,11 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
if url_data.valid:
|
||||
self.write(u"<tr><td bgcolor=\""+self.colorok+u"\">"+
|
||||
self.field("result")+u"</td><td bgcolor=\""+self.colorok+u"\">")
|
||||
self.part("result")+u"</td><td bgcolor=\""+self.colorok+u"\">")
|
||||
self.write(_("Valid"))
|
||||
else:
|
||||
self.write(u"<tr><td bgcolor=\""+self.colorerror+u"\">"+
|
||||
self.field("result")+u"</td><td bgcolor=\""+self.colorerror+u"\">")
|
||||
self.part("result")+u"</td><td bgcolor=\""+self.colorerror+u"\">")
|
||||
self.write(_("Error"))
|
||||
if url_data.result:
|
||||
self.write(u": "+cgi.escape(url_data.result))
|
||||
|
|
@ -280,7 +280,7 @@ class HtmlLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
if self.fd is None:
|
||||
return
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
self.writeln()
|
||||
self.write(_("That's it.")+" ")
|
||||
if self.number >= 0:
|
||||
|
|
|
|||
|
|
@ -77,7 +77,7 @@ class SQLLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.starttime = time.time()
|
||||
if self.has_field("intro"):
|
||||
if self.has_part("intro"):
|
||||
self.comment(_("created by %s at %s") % \
|
||||
(linkcheck.configuration.AppName,
|
||||
linkcheck.strformat.strtime(self.starttime)))
|
||||
|
|
@ -142,7 +142,7 @@ class SQLLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
if self.fd is None:
|
||||
return
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
self.stoptime = time.time()
|
||||
duration = self.stoptime - self.starttime
|
||||
self.comment(_("Stopped checking at %s (%s)") % \
|
||||
|
|
|
|||
|
|
@ -92,7 +92,7 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
if self.fd is None:
|
||||
return
|
||||
self.starttime = time.time()
|
||||
if self.has_field('intro'):
|
||||
if self.has_part('intro'):
|
||||
self.writeln(linkcheck.configuration.AppInfo)
|
||||
self.writeln(linkcheck.configuration.Freeware)
|
||||
self.writeln(_("Get the newest version at %s") %
|
||||
|
|
@ -111,27 +111,27 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
if self.fd is None:
|
||||
return
|
||||
if self.has_field('url'):
|
||||
if self.has_part('url'):
|
||||
self.write_url(url_data)
|
||||
if url_data.name and self.has_field('name'):
|
||||
if url_data.name and self.has_part('name'):
|
||||
self.write_name(url_data)
|
||||
if url_data.parent_url and self.has_field('parenturl'):
|
||||
if url_data.parent_url and self.has_part('parenturl'):
|
||||
self.write_parent(url_data)
|
||||
if url_data.base_ref and self.has_field('base'):
|
||||
if url_data.base_ref and self.has_part('base'):
|
||||
self.write_base(url_data)
|
||||
if url_data.url and self.has_field('realurl'):
|
||||
if url_data.url and self.has_part('realurl'):
|
||||
self.write_real(url_data)
|
||||
if url_data.dltime >= 0 and self.has_field('dltime'):
|
||||
if url_data.dltime >= 0 and self.has_part('dltime'):
|
||||
self.write_dltime(url_data)
|
||||
if url_data.dlsize >= 0 and self.has_field('dlsize'):
|
||||
if url_data.dlsize >= 0 and self.has_part('dlsize'):
|
||||
self.write_dlsize(url_data)
|
||||
if url_data.checktime and self.has_field('checktime'):
|
||||
if url_data.checktime and self.has_part('checktime'):
|
||||
self.write_checktime(url_data)
|
||||
if url_data.info and self.has_field('info'):
|
||||
if url_data.info and self.has_part('info'):
|
||||
self.write_info(url_data)
|
||||
if url_data.warning and self.has_field('warning'):
|
||||
if url_data.warning and self.has_part('warning'):
|
||||
self.write_warning(url_data)
|
||||
if self.has_field('result'):
|
||||
if self.has_part('result'):
|
||||
self.write_result(url_data)
|
||||
self.flush()
|
||||
|
||||
|
|
@ -140,7 +140,7 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
Write url_data.base_url.
|
||||
"""
|
||||
self.writeln()
|
||||
self.write(self.field('url') + self.spaces('url'))
|
||||
self.write(self.part('url') + self.spaces('url'))
|
||||
txt = unicode(repr(url_data.base_url)[1:])
|
||||
if url_data.cached:
|
||||
txt += _(" (cached)")
|
||||
|
|
@ -150,14 +150,14 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.name.
|
||||
"""
|
||||
self.write(self.field("name") + self.spaces("name"))
|
||||
self.write(self.part("name") + self.spaces("name"))
|
||||
self.writeln(unicode(repr(url_data.name)[1:]), color=self.colorname)
|
||||
|
||||
def write_parent (self, url_data):
|
||||
"""
|
||||
Write url_data.parent_url.
|
||||
"""
|
||||
self.write(self.field('parenturl') + self.spaces("parenturl"))
|
||||
self.write(self.part('parenturl') + self.spaces("parenturl"))
|
||||
txt = url_data.parent_url
|
||||
txt += _(", line %d") % url_data.line
|
||||
txt += _(", col %d") % url_data.column
|
||||
|
|
@ -167,21 +167,21 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.base_ref.
|
||||
"""
|
||||
self.write(self.field("base") + self.spaces("base"))
|
||||
self.write(self.part("base") + self.spaces("base"))
|
||||
self.writeln(url_data.base_ref, color=self.colorbase)
|
||||
|
||||
def write_real (self, url_data):
|
||||
"""
|
||||
Write url_data.url.
|
||||
"""
|
||||
self.write(self.field("realurl") + self.spaces("realurl"))
|
||||
self.write(self.part("realurl") + self.spaces("realurl"))
|
||||
self.writeln(unicode(url_data.url), color=self.colorreal)
|
||||
|
||||
def write_dltime (self, url_data):
|
||||
"""
|
||||
Write url_data.dltime.
|
||||
"""
|
||||
self.write(self.field("dltime") + self.spaces("dltime"))
|
||||
self.write(self.part("dltime") + self.spaces("dltime"))
|
||||
self.writeln(_("%.3f seconds") % url_data.dltime,
|
||||
color=self.colordltime)
|
||||
|
||||
|
|
@ -189,7 +189,7 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.dlsize.
|
||||
"""
|
||||
self.write(self.field("dlsize") + self.spaces("dlsize"))
|
||||
self.write(self.part("dlsize") + self.spaces("dlsize"))
|
||||
self.writeln(linkcheck.strformat.strsize(url_data.dlsize),
|
||||
color=self.colordlsize)
|
||||
|
||||
|
|
@ -197,7 +197,7 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.checktime.
|
||||
"""
|
||||
self.write(self.field("checktime") + self.spaces("checktime"))
|
||||
self.write(self.part("checktime") + self.spaces("checktime"))
|
||||
self.writeln(_("%.3f seconds") % url_data.checktime,
|
||||
color=self.colordltime)
|
||||
|
||||
|
|
@ -205,21 +205,21 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
Write url_data.info.
|
||||
"""
|
||||
self.write(self.field("info") + self.spaces("info"))
|
||||
self.write(self.part("info") + self.spaces("info"))
|
||||
self.writeln(self.wrap(url_data.info, 65), color=self.colorinfo)
|
||||
|
||||
def write_warning (self, url_data):
|
||||
"""
|
||||
Write url_data.warning.
|
||||
"""
|
||||
self.write(self.field("warning") + self.spaces("warning"))
|
||||
self.write(self.part("warning") + self.spaces("warning"))
|
||||
self.writeln(self.wrap(url_data.warning, 65), color=self.colorwarning)
|
||||
|
||||
def write_result (self, url_data):
|
||||
"""
|
||||
Write url_data.result.
|
||||
"""
|
||||
self.write(self.field("result") + self.spaces("result"))
|
||||
self.write(self.part("result") + self.spaces("result"))
|
||||
if url_data.valid:
|
||||
color = self.colorvalid
|
||||
self.write(_("Valid"), color=color)
|
||||
|
|
@ -236,7 +236,7 @@ class TextLogger (linkcheck.logger.Logger):
|
|||
"""
|
||||
if self.fd is None:
|
||||
return
|
||||
if self.has_field('outro'):
|
||||
if self.has_part('outro'):
|
||||
self.writeln()
|
||||
self.write(_("That's it.")+" ")
|
||||
if self.number >= 0:
|
||||
|
|
|
|||
|
|
@ -93,7 +93,7 @@ class XMLLogger (linkcheck.logger.Logger):
|
|||
return
|
||||
self.starttime = time.time()
|
||||
self.writeln(u'<?xml version="1.0"?>')
|
||||
if self.has_field("intro"):
|
||||
if self.has_part("intro"):
|
||||
self.comment(_("created by %s at %s") %
|
||||
(linkcheck.configuration.AppName,
|
||||
linkcheck.strformat.strtime(self.starttime)))
|
||||
|
|
@ -119,17 +119,17 @@ class XMLLogger (linkcheck.logger.Logger):
|
|||
self.nodes[node.url] = node
|
||||
self.nodeid += 1
|
||||
self.writeln(u' <node name="%d">' % node.id)
|
||||
if self.has_field("realurl"):
|
||||
if self.has_part("realurl"):
|
||||
self.writeln(u" <label>%s</label>" % xmlquote(node.url))
|
||||
self.writeln(u" <data>")
|
||||
if node.dltime >= 0 and self.has_field("dltime"):
|
||||
if node.dltime >= 0 and self.has_part("dltime"):
|
||||
self.writeln(u" <dltime>%f</dltime>" % node.dltime)
|
||||
if node.dlsize >= 0 and self.has_field("dlsize"):
|
||||
if node.dlsize >= 0 and self.has_part("dlsize"):
|
||||
self.writeln(u" <dlsize>%d</dlsize>" % node.dlsize)
|
||||
if node.checktime and self.has_field("checktime"):
|
||||
if node.checktime and self.has_part("checktime"):
|
||||
self.writeln(u" <checktime>%f</checktime>" %
|
||||
node.checktime)
|
||||
if self.has_field("extern"):
|
||||
if self.has_part("extern"):
|
||||
self.writeln(u" <extern>%d</extern>" %
|
||||
(node.extern and 1 or 0))
|
||||
self.writeln(u" </data>")
|
||||
|
|
@ -139,18 +139,18 @@ class XMLLogger (linkcheck.logger.Logger):
|
|||
def write_edges (self):
|
||||
"""
|
||||
Write all edges we can find in the graph in a brute-force
|
||||
manner. Better would be a mapping of parent urls.
|
||||
manner. Better would be a mapping of parent URLs.
|
||||
"""
|
||||
for node in self.nodes.values():
|
||||
if self.nodes.has_key(node.parent_url):
|
||||
self.write(u" <edge")
|
||||
self.write(u' source="%d"' % self.nodes[node.parent_url].id)
|
||||
self.writeln(u' target="%d">' % node.id)
|
||||
if self.has_field("url"):
|
||||
if self.has_part("url"):
|
||||
self.writeln(u" <label>%s</label>" % \
|
||||
xmlquote(node.base_url))
|
||||
self.writeln(u" <data>")
|
||||
if self.has_field("result"):
|
||||
if self.has_part("result"):
|
||||
self.writeln(u" <valid>%d</valid>" % \
|
||||
(node.valid and 1 or 0))
|
||||
self.writeln(u" </data>")
|
||||
|
|
@ -165,7 +165,7 @@ class XMLLogger (linkcheck.logger.Logger):
|
|||
return
|
||||
self.writeln(u"</graph>")
|
||||
self.writeln(u"</GraphXML>")
|
||||
if self.has_field("outro"):
|
||||
if self.has_part("outro"):
|
||||
self.stoptime = time.time()
|
||||
duration = self.stoptime - self.starttime
|
||||
self.comment(_("Stopped checking at %s (%s)") % \
|
||||
|
|
|
|||
Loading…
Reference in a new issue