diff --git a/linkcheck/logger/__init__.py b/linkcheck/logger/__init__.py index 61153769..f8912e36 100644 --- a/linkcheck/logger/__init__.py +++ b/linkcheck/logger/__init__.py @@ -51,15 +51,15 @@ class Logger (object): def __init__ (self, **args): """ - Initialize a logger, looking for field restrictions in kwargs. + Initialize a logger, looking for part restrictions in kwargs. """ - # what log fields should be in output - self.logfields = None # log all fields - if args.has_key('fields'): - if "all" not in args['fields']: - # only log given fields - self.logfields = args['fields'] - # number of spaces before log fields for alignment + # what log parts should be in output + self.logparts = None # log all parts + if args.has_key('parts'): + if "all" not in args['parts']: + # only log given parts + self.logparts = args['parts'] + # number of spaces before log parts for alignment self.logspaces = {} # maximum indent of spaces for alignment self.max_indent = 0 @@ -154,24 +154,24 @@ class Logger (object): self.write(s) self.write(unicode(os.linesep), **args) - def has_field (self, name): + def has_part (self, name): """ - See if given field name will be logged. + See if given part name will be logged. """ - if self.logfields is None: - # log all fields + if self.logparts is None: + # log all parts return True - return name in self.logfields + return name in self.logparts - def field (self, name): + def part (self, name): """ - Return translated field name. + Return translated part name. """ return _(Fields[name]) def spaces (self, name): """ - Return indent of spaces for given field name. + Return indent of spaces for given part name. """ return self.logspaces[name] @@ -179,16 +179,16 @@ class Logger (object): """ Start log output. """ - # map with spaces between field name and value - if self.logfields is None: - fields = Fields.keys() + # map with spaces between part name and value + if self.logparts is None: + parts = Fields.keys() else: - fields = self.logfields - values = [self.field(x) for x in fields] - # maximum indent for localized log field names + parts = self.logparts + values = [self.part(x) for x in parts] + # maximum indent for localized log part names self.max_indent = max([len(x) for x in values])+1 - for key in fields: - numspaces = (self.max_indent - len(self.field(key))) + for key in parts: + numspaces = (self.max_indent - len(self.part(key))) self.logspaces[key] = u" " * numspaces def log_filter_url (self, url_data, do_print): diff --git a/linkcheck/logger/csvlog.py b/linkcheck/logger/csvlog.py index 446fdf4e..c882c6ab 100644 --- a/linkcheck/logger/csvlog.py +++ b/linkcheck/logger/csvlog.py @@ -57,7 +57,7 @@ class CSVLogger (linkcheck.logger.Logger): return self.starttime = time.time() row = [] - if self.has_field("intro"): + if self.has_part("intro"): self.comment(_("created by %s at %s") % \ (linkcheck.configuration.AppName, linkcheck.strformat.strtime(self.starttime))) @@ -123,7 +123,7 @@ class CSVLogger (linkcheck.logger.Logger): if self.fd is None: return self.stoptime = time.time() - if self.has_field("outro"): + if self.has_part("outro"): duration = self.stoptime - self.starttime self.comment(_("Stopped checking at %s (%s)") % \ (linkcheck.strformat.strtime(self.stoptime), diff --git a/linkcheck/logger/dot.py b/linkcheck/logger/dot.py index ef83152e..670c015a 100644 --- a/linkcheck/logger/dot.py +++ b/linkcheck/logger/dot.py @@ -46,7 +46,7 @@ class DOTLogger (linkcheck.logger.Logger): if self.fd is None: return self.starttime = time.time() - if self.has_field("intro"): + if self.has_part("intro"): self.comment(_("created by %s at %s") % \ (linkcheck.configuration.AppName, linkcheck.strformat.strtime(self.starttime))) @@ -78,22 +78,22 @@ class DOTLogger (linkcheck.logger.Logger): self.nodes[node.url] = node self.nodeid += 1 self.writeln(u" %d [" % node.id) - if self.has_field("realurl"): + if self.has_part("realurl"): self.writeln(u' href="%s",' % dotquote(node.url)) - if node.dltime >= 0 and self.has_field("dltime"): + if node.dltime >= 0 and self.has_part("dltime"): self.writeln(u" dltime=%d," % node.dltime) - if node.dlsize >= 0 and self.has_field("dlsize"): + if node.dlsize >= 0 and self.has_part("dlsize"): self.writeln(u" dlsize=%d," % node.dlsize) - if node.checktime and self.has_field("checktime"): + if node.checktime and self.has_part("checktime"): self.writeln(u" checktime=%d," % node.checktime) - if self.has_field("extern"): + if self.has_part("extern"): self.writeln(u" extern=%d," % (node.extern and 1 or 0)) self.writeln(u" ];") def write_edges (self): """ Write all edges we can find in the graph in a brute-force - manner. Better would be a mapping of parent urls. + manner. Better would be a mapping of parent URLs. """ for node in self.nodes.values(): if self.nodes.has_key(node.parent_url): @@ -101,7 +101,7 @@ class DOTLogger (linkcheck.logger.Logger): target = node.id self.writeln(u" %d -> %d [" % (source, target)) self.writeln(u' label="%s",' % dotedge(node.name)) - if self.has_field("result"): + if self.has_part("result"): self.writeln(u" valid=%d," % (node.valid and 1 or 0)) self.writeln(u" ];") self.flush() @@ -114,7 +114,7 @@ class DOTLogger (linkcheck.logger.Logger): return self.write_edges() self.writeln(u"}") - if self.has_field("outro"): + if self.has_part("outro"): self.stoptime = time.time() duration = self.stoptime - self.starttime self.comment(_("Stopped checking at %s (%s)")%\ diff --git a/linkcheck/logger/gml.py b/linkcheck/logger/gml.py index 3f84611c..a8bf84be 100644 --- a/linkcheck/logger/gml.py +++ b/linkcheck/logger/gml.py @@ -46,7 +46,7 @@ class GMLLogger (linkcheck.logger.Logger): if self.fd is None: return self.starttime = time.time() - if self.has_field("intro"): + if self.has_part("intro"): self.comment(_("created by %s at %s") % \ (linkcheck.configuration.AppName, linkcheck.strformat.strtime(self.starttime))) @@ -80,32 +80,32 @@ class GMLLogger (linkcheck.logger.Logger): self.nodeid += 1 self.writeln(u" node [") self.writeln(u" id %d" % node.id) - if self.has_field("realurl"): + if self.has_part("realurl"): self.writeln(u' label "%s"' % node.url) - if node.dltime >= 0 and self.has_field("dltime"): + if node.dltime >= 0 and self.has_part("dltime"): self.writeln(u" dltime %d" % node.dltime) - if node.dlsize >= 0 and self.has_field("dlsize"): + if node.dlsize >= 0 and self.has_part("dlsize"): self.writeln(u" dlsize %d" % node.dlsize) - if node.checktime and self.has_field("checktime"): + if node.checktime and self.has_part("checktime"): self.writeln(u" checktime %d" % node.checktime) - if self.has_field("extern"): + if self.has_part("extern"): self.writeln(u" extern %d" % (node.extern and 1 or 0)) self.writeln(u" ]") def write_edges (self): """ Write all edges we can find in the graph in a brute-force - manner. Better would be a mapping of parent urls. + manner. Better would be a mapping of parent URLs. """ for node in self.nodes.values(): if self.nodes.has_key(node.parent_url): self.writeln(u" edge [") self.writeln(u' label "%s"' % node.base_url) - if self.has_field("parenturl"): + if self.has_part("parenturl"): self.writeln(u" source %d" % \ self.nodes[node.parent_url].id) self.writeln(u" target %d" % node.id) - if self.has_field("result"): + if self.has_part("result"): self.writeln(u" valid %d" % (node.valid and 1 or 0)) self.writeln(u" ]") self.flush() @@ -118,7 +118,7 @@ class GMLLogger (linkcheck.logger.Logger): return self.write_edges() self.writeln(u"]") - if self.has_field("outro"): + if self.has_part("outro"): self.stoptime = time.time() duration = self.stoptime - self.starttime self.comment(_("Stopped checking at %s (%s)")%\ diff --git a/linkcheck/logger/html.py b/linkcheck/logger/html.py index 2e55f657..df45ccd1 100644 --- a/linkcheck/logger/html.py +++ b/linkcheck/logger/html.py @@ -72,11 +72,11 @@ class HtmlLogger (linkcheck.logger.Logger): self.colorerror = args['colorerror'] self.colorok = args['colorok'] - def field (self, name): + def part (self, name): """ - Return non-space-breakable field name. + Return non-space-breakable part name. """ - return super(HtmlLogger, self).field(name).replace(" ", " ") + return super(HtmlLogger, self).part(name).replace(" ", " ") def comment (self, s, **args): """ @@ -97,7 +97,7 @@ class HtmlLogger (linkcheck.logger.Logger): self.write(HTML_HEADER % (linkcheck.configuration.App, self.colorbackground, self.colorlink, self.colorlink, self.colorlink)) - if self.has_field('intro'): + if self.has_part('intro'): self.write(u"

"+linkcheck.configuration.App+ "


"+ linkcheck.configuration.Freeware+"

"+ @@ -114,27 +114,27 @@ class HtmlLogger (linkcheck.logger.Logger): if self.fd is None: return self.write_table_start() - if self.has_field("url"): + if self.has_part("url"): self.write_url(url_data) - if url_data.name and self.has_field("name"): + if url_data.name and self.has_part("name"): self.write_name(url_data) - if url_data.parent_url and self.has_field("parenturl"): + if url_data.parent_url and self.has_part("parenturl"): self.write_parent(url_data) - if url_data.base_ref and self.has_field("base"): + if url_data.base_ref and self.has_part("base"): self.write_base(url_data) - if url_data.url and self.has_field("realurl"): + if url_data.url and self.has_part("realurl"): self.write_real(url_data) - if url_data.dltime >= 0 and self.has_field("dltime"): + if url_data.dltime >= 0 and self.has_part("dltime"): self.write_dltime(url_data) - if url_data.dlsize >= 0 and self.has_field("dlsize"): + if url_data.dlsize >= 0 and self.has_part("dlsize"): self.write_dlsize(url_data) - if url_data.checktime and self.has_field("checktime"): + if url_data.checktime and self.has_part("checktime"): self.write_checktime(url_data) - if url_data.info and self.has_field("info"): + if url_data.info and self.has_part("info"): self.write_info(url_data) - if url_data.warning and self.has_field("warning"): + if url_data.warning and self.has_part("warning"): self.write_warning(url_data) - if self.has_field("result"): + if self.has_part("result"): self.write_result(url_data) self.write_table_end() self.flush() @@ -167,7 +167,7 @@ class HtmlLogger (linkcheck.logger.Logger): """ self.writeln(u"") self.writeln(u""+ - self.field("url")+u"") + self.part("url")+u"") self.write(u""+ cgi.escape(repr(url_data.base_url)[1:])) if url_data.cached: @@ -178,14 +178,14 @@ class HtmlLogger (linkcheck.logger.Logger): """ Write url_data.name. """ - self.writeln(u""+self.field("name")+u""+ + self.writeln(u""+self.part("name")+u""+ cgi.escape(repr(url_data.name)[1:])+u"") def write_parent (self, url_data): """ Write url_data.parent_url. """ - self.write(u""+self.field("parenturl")+ + self.write(u""+self.part("parenturl")+ u''+ cgi.escape(url_data.parent_url)+u"") @@ -204,14 +204,14 @@ class HtmlLogger (linkcheck.logger.Logger): """ Write url_data.base_ref. """ - self.writeln(u""+self.field("base")+u""+ + self.writeln(u""+self.part("base")+u""+ cgi.escape(url_data.base_ref)+u"") def write_real (self, url_data): """ Write url_data.url. """ - self.writeln(""+self.field("realurl")+u""+ + self.writeln(""+self.part("realurl")+u""+ u''+cgi.escape(url_data.url)+u"") @@ -219,7 +219,7 @@ class HtmlLogger (linkcheck.logger.Logger): """ Write url_data.dltime. """ - self.writeln(u""+self.field("dltime")+u""+ + self.writeln(u""+self.part("dltime")+u""+ (_("%.3f seconds") % url_data.dltime)+ u"") @@ -227,7 +227,7 @@ class HtmlLogger (linkcheck.logger.Logger): """ Write url_data.dlsize. """ - self.writeln(u""+self.field("dlsize")+u""+ + self.writeln(u""+self.part("dlsize")+u""+ linkcheck.strformat.strsize(url_data.dlsize)+ u"") @@ -235,7 +235,7 @@ class HtmlLogger (linkcheck.logger.Logger): """ Write url_data.checktime. """ - self.writeln(u""+self.field("checktime")+u""+ + self.writeln(u""+self.part("checktime")+u""+ (_("%.3f seconds") % url_data.checktime)+u"") def write_info (self, url_data): @@ -243,7 +243,7 @@ class HtmlLogger (linkcheck.logger.Logger): Write url_data.info. """ text = os.linesep.join(url_data.info) - self.writeln(u""+self.field("info")+ + self.writeln(u""+self.part("info")+ u""+cgi.escape(text).replace(os.linesep, "
")+ u"") @@ -254,7 +254,7 @@ class HtmlLogger (linkcheck.logger.Logger): sep = u"
"+os.linesep text = sep.join([cgi.escape(x) for x in url_data.warning]) self.writeln(u""+self.field("warning")+ + u"valign=\"top\">"+self.part("warning")+ u""+ text+u"") @@ -264,11 +264,11 @@ class HtmlLogger (linkcheck.logger.Logger): """ if url_data.valid: self.write(u""+ - self.field("result")+u"") + self.part("result")+u"") self.write(_("Valid")) else: self.write(u""+ - self.field("result")+u"") + self.part("result")+u"") self.write(_("Error")) if url_data.result: self.write(u": "+cgi.escape(url_data.result)) @@ -280,7 +280,7 @@ class HtmlLogger (linkcheck.logger.Logger): """ if self.fd is None: return - if self.has_field("outro"): + if self.has_part("outro"): self.writeln() self.write(_("That's it.")+" ") if self.number >= 0: diff --git a/linkcheck/logger/sql.py b/linkcheck/logger/sql.py index 83df49e1..20426872 100644 --- a/linkcheck/logger/sql.py +++ b/linkcheck/logger/sql.py @@ -77,7 +77,7 @@ class SQLLogger (linkcheck.logger.Logger): if self.fd is None: return self.starttime = time.time() - if self.has_field("intro"): + if self.has_part("intro"): self.comment(_("created by %s at %s") % \ (linkcheck.configuration.AppName, linkcheck.strformat.strtime(self.starttime))) @@ -142,7 +142,7 @@ class SQLLogger (linkcheck.logger.Logger): """ if self.fd is None: return - if self.has_field("outro"): + if self.has_part("outro"): self.stoptime = time.time() duration = self.stoptime - self.starttime self.comment(_("Stopped checking at %s (%s)") % \ diff --git a/linkcheck/logger/text.py b/linkcheck/logger/text.py index 1cf76475..012e4ee7 100644 --- a/linkcheck/logger/text.py +++ b/linkcheck/logger/text.py @@ -92,7 +92,7 @@ class TextLogger (linkcheck.logger.Logger): if self.fd is None: return self.starttime = time.time() - if self.has_field('intro'): + if self.has_part('intro'): self.writeln(linkcheck.configuration.AppInfo) self.writeln(linkcheck.configuration.Freeware) self.writeln(_("Get the newest version at %s") % @@ -111,27 +111,27 @@ class TextLogger (linkcheck.logger.Logger): """ if self.fd is None: return - if self.has_field('url'): + if self.has_part('url'): self.write_url(url_data) - if url_data.name and self.has_field('name'): + if url_data.name and self.has_part('name'): self.write_name(url_data) - if url_data.parent_url and self.has_field('parenturl'): + if url_data.parent_url and self.has_part('parenturl'): self.write_parent(url_data) - if url_data.base_ref and self.has_field('base'): + if url_data.base_ref and self.has_part('base'): self.write_base(url_data) - if url_data.url and self.has_field('realurl'): + if url_data.url and self.has_part('realurl'): self.write_real(url_data) - if url_data.dltime >= 0 and self.has_field('dltime'): + if url_data.dltime >= 0 and self.has_part('dltime'): self.write_dltime(url_data) - if url_data.dlsize >= 0 and self.has_field('dlsize'): + if url_data.dlsize >= 0 and self.has_part('dlsize'): self.write_dlsize(url_data) - if url_data.checktime and self.has_field('checktime'): + if url_data.checktime and self.has_part('checktime'): self.write_checktime(url_data) - if url_data.info and self.has_field('info'): + if url_data.info and self.has_part('info'): self.write_info(url_data) - if url_data.warning and self.has_field('warning'): + if url_data.warning and self.has_part('warning'): self.write_warning(url_data) - if self.has_field('result'): + if self.has_part('result'): self.write_result(url_data) self.flush() @@ -140,7 +140,7 @@ class TextLogger (linkcheck.logger.Logger): Write url_data.base_url. """ self.writeln() - self.write(self.field('url') + self.spaces('url')) + self.write(self.part('url') + self.spaces('url')) txt = unicode(repr(url_data.base_url)[1:]) if url_data.cached: txt += _(" (cached)") @@ -150,14 +150,14 @@ class TextLogger (linkcheck.logger.Logger): """ Write url_data.name. """ - self.write(self.field("name") + self.spaces("name")) + self.write(self.part("name") + self.spaces("name")) self.writeln(unicode(repr(url_data.name)[1:]), color=self.colorname) def write_parent (self, url_data): """ Write url_data.parent_url. """ - self.write(self.field('parenturl') + self.spaces("parenturl")) + self.write(self.part('parenturl') + self.spaces("parenturl")) txt = url_data.parent_url txt += _(", line %d") % url_data.line txt += _(", col %d") % url_data.column @@ -167,21 +167,21 @@ class TextLogger (linkcheck.logger.Logger): """ Write url_data.base_ref. """ - self.write(self.field("base") + self.spaces("base")) + self.write(self.part("base") + self.spaces("base")) self.writeln(url_data.base_ref, color=self.colorbase) def write_real (self, url_data): """ Write url_data.url. """ - self.write(self.field("realurl") + self.spaces("realurl")) + self.write(self.part("realurl") + self.spaces("realurl")) self.writeln(unicode(url_data.url), color=self.colorreal) def write_dltime (self, url_data): """ Write url_data.dltime. """ - self.write(self.field("dltime") + self.spaces("dltime")) + self.write(self.part("dltime") + self.spaces("dltime")) self.writeln(_("%.3f seconds") % url_data.dltime, color=self.colordltime) @@ -189,7 +189,7 @@ class TextLogger (linkcheck.logger.Logger): """ Write url_data.dlsize. """ - self.write(self.field("dlsize") + self.spaces("dlsize")) + self.write(self.part("dlsize") + self.spaces("dlsize")) self.writeln(linkcheck.strformat.strsize(url_data.dlsize), color=self.colordlsize) @@ -197,7 +197,7 @@ class TextLogger (linkcheck.logger.Logger): """ Write url_data.checktime. """ - self.write(self.field("checktime") + self.spaces("checktime")) + self.write(self.part("checktime") + self.spaces("checktime")) self.writeln(_("%.3f seconds") % url_data.checktime, color=self.colordltime) @@ -205,21 +205,21 @@ class TextLogger (linkcheck.logger.Logger): """ Write url_data.info. """ - self.write(self.field("info") + self.spaces("info")) + self.write(self.part("info") + self.spaces("info")) self.writeln(self.wrap(url_data.info, 65), color=self.colorinfo) def write_warning (self, url_data): """ Write url_data.warning. """ - self.write(self.field("warning") + self.spaces("warning")) + self.write(self.part("warning") + self.spaces("warning")) self.writeln(self.wrap(url_data.warning, 65), color=self.colorwarning) def write_result (self, url_data): """ Write url_data.result. """ - self.write(self.field("result") + self.spaces("result")) + self.write(self.part("result") + self.spaces("result")) if url_data.valid: color = self.colorvalid self.write(_("Valid"), color=color) @@ -236,7 +236,7 @@ class TextLogger (linkcheck.logger.Logger): """ if self.fd is None: return - if self.has_field('outro'): + if self.has_part('outro'): self.writeln() self.write(_("That's it.")+" ") if self.number >= 0: diff --git a/linkcheck/logger/xmllog.py b/linkcheck/logger/xmllog.py index 04f44e95..89c142f9 100644 --- a/linkcheck/logger/xmllog.py +++ b/linkcheck/logger/xmllog.py @@ -93,7 +93,7 @@ class XMLLogger (linkcheck.logger.Logger): return self.starttime = time.time() self.writeln(u'') - if self.has_field("intro"): + if self.has_part("intro"): self.comment(_("created by %s at %s") % (linkcheck.configuration.AppName, linkcheck.strformat.strtime(self.starttime))) @@ -119,17 +119,17 @@ class XMLLogger (linkcheck.logger.Logger): self.nodes[node.url] = node self.nodeid += 1 self.writeln(u' ' % node.id) - if self.has_field("realurl"): + if self.has_part("realurl"): self.writeln(u" " % xmlquote(node.url)) self.writeln(u" ") - if node.dltime >= 0 and self.has_field("dltime"): + if node.dltime >= 0 and self.has_part("dltime"): self.writeln(u" %f" % node.dltime) - if node.dlsize >= 0 and self.has_field("dlsize"): + if node.dlsize >= 0 and self.has_part("dlsize"): self.writeln(u" %d" % node.dlsize) - if node.checktime and self.has_field("checktime"): + if node.checktime and self.has_part("checktime"): self.writeln(u" %f" % node.checktime) - if self.has_field("extern"): + if self.has_part("extern"): self.writeln(u" %d" % (node.extern and 1 or 0)) self.writeln(u" ") @@ -139,18 +139,18 @@ class XMLLogger (linkcheck.logger.Logger): def write_edges (self): """ Write all edges we can find in the graph in a brute-force - manner. Better would be a mapping of parent urls. + manner. Better would be a mapping of parent URLs. """ for node in self.nodes.values(): if self.nodes.has_key(node.parent_url): self.write(u" ' % node.id) - if self.has_field("url"): + if self.has_part("url"): self.writeln(u" " % \ xmlquote(node.base_url)) self.writeln(u" ") - if self.has_field("result"): + if self.has_part("result"): self.writeln(u" %d" % \ (node.valid and 1 or 0)) self.writeln(u" ") @@ -165,7 +165,7 @@ class XMLLogger (linkcheck.logger.Logger): return self.writeln(u"") self.writeln(u"") - if self.has_field("outro"): + if self.has_part("outro"): self.stoptime = time.time() duration = self.stoptime - self.starttime self.comment(_("Stopped checking at %s (%s)") % \