linkchecker/linkcheck/logger/gml.py
2005-07-16 07:12:44 +00:00

127 lines
4.6 KiB
Python

# -*- coding: iso-8859-1 -*-
# Copyright (C) 2000-2005 Bastian Kleineidam
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
"""
A gml logger.
"""
import time
import linkcheck.configuration
class GMLLogger (linkcheck.logger.Logger):
"""
GML means Graph Modeling Language. Use a GML tool to see
the sitemap graph.
"""
def __init__ (self, **args):
"""
Initialize graph node list and internal id counter.
"""
super(GMLLogger, self).__init__(**args)
self.init_fileoutput(args)
self.nodes = {}
self.nodeid = 0
def start_output (self):
"""
Write start of checking info as gml comment.
"""
super(GMLLogger, self).start_output()
if self.fd is None:
return
self.starttime = time.time()
if self.has_part("intro"):
self.comment(_("created by %s at %s") % \
(linkcheck.configuration.AppName,
linkcheck.strformat.strtime(self.starttime)))
self.comment(_("Get the newest version at %(url)s") % \
{'url': linkcheck.configuration.Url})
self.comment(_("Write comments and bugs to %(email)s") % \
{'email': linkcheck.configuration.Email})
self.check_date()
self.writeln()
self.writeln(u"graph [")
self.writeln(u" directed 1")
self.flush()
def comment (self, s, **args):
"""
Write GML comment.
"""
self.write(u"# ")
self.writeln(s=s, **args)
def log_url (self, url_data):
"""
Write one node and all possible edges.
"""
if self.fd is None:
return
node = url_data
if node.url and not self.nodes.has_key(node.url):
node.id = self.nodeid
self.nodes[node.url] = node
self.nodeid += 1
self.writeln(u" node [")
self.writeln(u" id %d" % node.id)
if self.has_part("realurl"):
self.writeln(u' label "%s"' % node.url)
if node.dltime >= 0 and self.has_part("dltime"):
self.writeln(u" dltime %d" % node.dltime)
if node.dlsize >= 0 and self.has_part("dlsize"):
self.writeln(u" dlsize %d" % node.dlsize)
if node.checktime and self.has_part("checktime"):
self.writeln(u" checktime %d" % node.checktime)
if self.has_part("extern"):
self.writeln(u" extern %d" % (node.extern[0] and 1 or 0))
self.writeln(u" ]")
def write_edges (self):
"""
Write all edges we can find in the graph in a brute-force
manner. Better would be a mapping of parent URLs.
"""
for node in self.nodes.values():
if self.nodes.has_key(node.parent_url):
self.writeln(u" edge [")
self.writeln(u' label "%s"' % (node.base_url or u""))
if self.has_part("parenturl") and node.parent_url:
self.writeln(u" source %d" % \
self.nodes[node.parent_url].id)
self.writeln(u" target %d" % node.id)
if self.has_part("result"):
self.writeln(u" valid %d" % (node.valid and 1 or 0))
self.writeln(u" ]")
self.flush()
def end_output (self):
"""
Write end of checking info as gml comment.
"""
if self.fd is None:
return
self.write_edges()
self.writeln(u"]")
if self.has_part("outro"):
self.stoptime = time.time()
duration = self.stoptime - self.starttime
self.comment(_("Stopped checking at %s (%s)")%\
(linkcheck.strformat.strtime(self.stoptime),
linkcheck.strformat.strduration(duration)))
self.close_fileoutput()