mirror of
https://github.com/Hopiu/linkchecker.git
synced 2026-03-21 08:20:25 +00:00
git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@2900 e7d03fd6-7b0d-0410-9947-9c21f3af8025
387 lines
13 KiB
Python
387 lines
13 KiB
Python
# -*- coding: iso-8859-1 -*-
|
|
# Copyright (C) 2000-2005 Bastian Kleineidam
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation; either version 2 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program; if not, write to the Free Software
|
|
# Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
|
|
"""
|
|
Main functions for link checking.
|
|
"""
|
|
|
|
import time
|
|
import sys
|
|
import os
|
|
import cgi
|
|
import socket
|
|
import codecs
|
|
import traceback
|
|
import select
|
|
import re
|
|
import urllib
|
|
import nntplib
|
|
import ftplib
|
|
|
|
import linkcheck.httplib2
|
|
import linkcheck.strformat
|
|
import linkcheck.dns.exception
|
|
|
|
|
|
# Catch these exception on syntax checks.
|
|
ExcSyntaxList = [
|
|
linkcheck.LinkCheckerError,
|
|
# .encode('idna') raises this
|
|
UnicodeError,
|
|
]
|
|
|
|
# Catch these exceptions on content and connect checks. All other
|
|
# exceptions are internal or system errors
|
|
ExcCacheList = [
|
|
IOError,
|
|
OSError, # OSError is thrown on Windows when a file is not found
|
|
linkcheck.LinkCheckerError,
|
|
linkcheck.dns.exception.DNSException,
|
|
socket.error,
|
|
select.error,
|
|
# nttp errors (including EOFError)
|
|
nntplib.error_reply,
|
|
nntplib.error_temp,
|
|
nntplib.error_perm,
|
|
nntplib.error_proto,
|
|
EOFError,
|
|
# http error
|
|
linkcheck.httplib2.error,
|
|
# ftp errors
|
|
ftplib.error_reply,
|
|
ftplib.error_temp,
|
|
ftplib.error_perm,
|
|
ftplib.error_proto,
|
|
]
|
|
|
|
# Exceptions that do not put the URL in the cache so that the URL can
|
|
# be checked again.
|
|
ExcNoCacheList = [
|
|
socket.timeout,
|
|
]
|
|
|
|
ExcList = ExcCacheList + ExcNoCacheList
|
|
|
|
# registered warnings
|
|
Warnings = {
|
|
"url-effective-url":
|
|
_("The effective URL is different from the original."),
|
|
"url-error-getting-content":
|
|
_("Could not get the content of the URL."),
|
|
"url-unicode-domain": _("URL uses a unicode domain."),
|
|
"url-unnormed": _("URL is not normed."),
|
|
"url-anchor-not-found": _("URL anchor was not found."),
|
|
"url-warnregex-found":
|
|
_("The warning regular expression was found in the URL contents."),
|
|
"url-content-too-large": _("The URL content is too large."),
|
|
"file-missing-slash": _("The file: URL is missing a trailing slash."),
|
|
"file-system-path":
|
|
_("The file: path is not the same as the system specific path."),
|
|
"ftp-missing-slash": _("The ftp: URL is missing a trailing slash."),
|
|
"http-robots-denied": _("The http: URL checking has been denied."),
|
|
"http-no-anchor-support": _("The HTTP server had no anchor support."),
|
|
"http-moved-permanent": _("The URL has moved permanently."),
|
|
"http-wrong-redirect":
|
|
_("The URL has been redirected to an URL of a different type."),
|
|
"http-empty-content": _("The URL had no content."),
|
|
"http-cookie-store-error": _("An error occurred while storing a cookie."),
|
|
"http-decompress-error":
|
|
_("An error occurred while decompressing the URL content."),
|
|
"http-unsupported-encoding":
|
|
_("The URL content is encoded with an unknown encoding."),
|
|
"ignored-url": _("The URL has been ignored."),
|
|
"mail-no-addresses": _("The mailto: URL contained no addresses."),
|
|
"mail-no-mx-host": _("The mail MX host could not be found."),
|
|
"mail-unverified-address":
|
|
_("The mailto: address could not be verified."),
|
|
"mail-no-connection":
|
|
_("No connection to a MX host could be established."),
|
|
"nntp-no-server": _("No NNTP server was found."),
|
|
"nntp-no-newsgroup": _("The NNTP newsgroup could not be found."),
|
|
"nntp-busy": _("The NNTP server was busy."),
|
|
}
|
|
|
|
ignored_schemes = r"""^(
|
|
acap # application configuration access protocol
|
|
|afs # Andrew File System global file names
|
|
|cid # content identifier
|
|
|data # data
|
|
|dav # dav
|
|
|fax # fax
|
|
|imap # internet message access protocol
|
|
|ldap # Lightweight Directory Access Protocol
|
|
|mailserver # Access to data available from mail servers
|
|
|mid # message identifier
|
|
|mms # multimedia stream
|
|
|modem # modem
|
|
|nfs # network file system protocol
|
|
|opaquelocktoken # opaquelocktoken
|
|
|pop # Post Office Protocol v3
|
|
|prospero # Prospero Directory Service
|
|
|rsync # rsync protocol
|
|
|rtsp # real time streaming protocol
|
|
|rtspu # real time streaming protocol
|
|
|service # service location
|
|
|shttp # secure HTTP
|
|
|sip # session initiation protocol
|
|
|tel # telephone
|
|
|tip # Transaction Internet Protocol
|
|
|tn3270 # Interactive 3270 emulation sessions
|
|
|vemmi # versatile multimedia interface
|
|
|wais # Wide Area Information Servers
|
|
|z39\.50r # Z39.50 Retrieval
|
|
|z39\.50s # Z39.50 Session
|
|
|chrome # Mozilla specific
|
|
|find # Mozilla specific
|
|
|clsid # Microsoft specific
|
|
|javascript # JavaScript
|
|
|isbn # ISBN (int. book numbers)
|
|
):"""
|
|
|
|
ignored_schemes_re = re.compile(ignored_schemes, re.VERBOSE)
|
|
|
|
_encoding = linkcheck.i18n.default_encoding
|
|
stderr = codecs.getwriter(_encoding)(sys.stderr, errors="ignore")
|
|
|
|
def internal_error ():
|
|
"""
|
|
Print internal error message to stderr.
|
|
"""
|
|
print >> stderr, os.linesep
|
|
print >> stderr, _("""********** Oops, I did it again. *************
|
|
|
|
You have found an internal error in LinkChecker. Please write a bug report
|
|
at http://sourceforge.net/tracker/?func=add&group_id=1913&atid=101913
|
|
or send mail to %s and include the following information:
|
|
- the URL or file you are testing
|
|
- your commandline arguments and/or configuration.
|
|
- the output of a debug run with option "-Dall" of the executed command
|
|
- the system information below.
|
|
|
|
Disclosing some of the information above due to privacy reasons is ok.
|
|
I will try to help you nonetheless, but you have to give me something
|
|
I can work with ;) .
|
|
""") % linkcheck.configuration.Email
|
|
etype, value = sys.exc_info()[:2]
|
|
print >> stderr, etype, value
|
|
traceback.print_exc()
|
|
print_app_info()
|
|
print >> stderr, os.linesep, \
|
|
_("******** LinkChecker internal error, over and out ********")
|
|
sys.exit(1)
|
|
|
|
|
|
def print_app_info ():
|
|
"""
|
|
Print system and application info to stderr.
|
|
"""
|
|
print >> stderr, _("System info:")
|
|
print >> stderr, linkcheck.configuration.App
|
|
print >> stderr, _("Python %s on %s") % (sys.version, sys.platform)
|
|
for key in ("LC_ALL", "LC_MESSAGES", "http_proxy", "ftp_proxy"):
|
|
value = os.getenv(key)
|
|
if value is not None:
|
|
print >> stderr, key, "=", repr(value)
|
|
|
|
|
|
def check_urls (consumer):
|
|
"""
|
|
Main check function; checks all configured URLs until interrupted
|
|
with Ctrl-C. If you call this function more than once, you can specify
|
|
different configurations with the consumer parameter.
|
|
|
|
@param consumer: an object where all runtime-dependent options are
|
|
stored
|
|
@type consumer: linkcheck.consumer.Consumer
|
|
@return: None
|
|
"""
|
|
try:
|
|
_check_urls(consumer)
|
|
except (KeyboardInterrupt, SystemExit):
|
|
consumer.abort()
|
|
except:
|
|
consumer.abort()
|
|
internal_error()
|
|
|
|
|
|
def _check_urls (consumer):
|
|
"""
|
|
Checks all configured URLs. Prints status information, calls logger
|
|
methods.
|
|
|
|
@param consumer: an object where all runtime-dependent options are
|
|
stored
|
|
@type consumer: linkcheck.consumer.Consumer
|
|
@return: None
|
|
"""
|
|
start_time = time.time()
|
|
status_time = start_time
|
|
while not consumer.finished():
|
|
url_data = consumer.incoming_get_url()
|
|
if url_data is None:
|
|
# wait for incoming queue to fill
|
|
time.sleep(0.1)
|
|
elif url_data.cached:
|
|
# was cached -> can be logged
|
|
consumer.log_url(url_data)
|
|
else:
|
|
# go check this url
|
|
if url_data.parent_url and not \
|
|
linkcheck.url.url_is_absolute(url_data.base_url):
|
|
name = url_data.parent_url
|
|
else:
|
|
name = u""
|
|
if url_data.base_url:
|
|
name += url_data.base_url
|
|
if not name:
|
|
name = None
|
|
consumer.check_url(url_data, name)
|
|
if consumer.config('status'):
|
|
curtime = time.time()
|
|
if (curtime - status_time) > 5:
|
|
consumer.print_status(curtime, start_time)
|
|
status_time = curtime
|
|
consumer.end_log_output()
|
|
|
|
|
|
# file extensions we can parse recursively
|
|
extensions = {
|
|
"html": re.compile(r'(?i)\.s?html?$'),
|
|
"opera": re.compile(r'^(?i)opera.adr$'), # opera bookmark file
|
|
"css": re.compile(r'(?i)\.css$'), # CSS stylesheet
|
|
}
|
|
|
|
|
|
import linkcheck.checker.fileurl
|
|
import linkcheck.checker.ignoredurl
|
|
import linkcheck.checker.ftpurl
|
|
import linkcheck.checker.gopherurl
|
|
import linkcheck.checker.httpurl
|
|
import linkcheck.checker.httpsurl
|
|
import linkcheck.checker.mailtourl
|
|
import linkcheck.checker.telneturl
|
|
import linkcheck.checker.nntpurl
|
|
import linkcheck.checker.errorurl
|
|
|
|
|
|
def absolute_url (base_url, base_ref, parent_url):
|
|
"""
|
|
Search for the absolute url to detect the link type. This does not
|
|
join any url fragments together! Returns the url in lower case to
|
|
simplify urltype matching.
|
|
|
|
@param base_url: base url from a link tag
|
|
@type base_url: string or None
|
|
@param base_ref: base url from <base> tag
|
|
@type base_ref: string or None
|
|
@param parent_url: url of parent document
|
|
@type parent_url: string or None
|
|
"""
|
|
if base_url and linkcheck.url.url_is_absolute(base_url):
|
|
return base_url.lower()
|
|
elif base_ref and linkcheck.url.url_is_absolute(base_ref):
|
|
return base_ref.lower()
|
|
elif parent_url and linkcheck.url.url_is_absolute(parent_url):
|
|
return parent_url.lower()
|
|
return u""
|
|
|
|
|
|
def get_url_from (base_url, recursion_level, consumer,
|
|
parent_url=None, base_ref=None, line=0, column=0,
|
|
name=u"", cmdline=False):
|
|
"""
|
|
Get url data from given base data.
|
|
|
|
@param base_url: base url from a link tag
|
|
@type base_url: string or None
|
|
@param recursion_level: current recursion level
|
|
@type recursion_level: number
|
|
@param consumer: consumer object
|
|
@type consumer: linkcheck.checker.consumer.Consumer
|
|
@param parent_url: parent url
|
|
@type parent_url: string or None
|
|
@param base_ref: base url from <base> tag
|
|
@type base_ref string or None
|
|
@param line: line number
|
|
@type line: number
|
|
@param column: column number
|
|
@type column: number
|
|
@param name: link name
|
|
@type name: string
|
|
"""
|
|
if base_url is not None:
|
|
base_url = linkcheck.strformat.unicode_safe(base_url)
|
|
if parent_url is not None:
|
|
parent_url = linkcheck.strformat.unicode_safe(parent_url)
|
|
if base_ref is not None:
|
|
base_ref = linkcheck.strformat.unicode_safe(base_ref)
|
|
name = linkcheck.strformat.unicode_safe(name)
|
|
url = absolute_url(base_url, base_ref, parent_url)
|
|
# test scheme
|
|
if url.startswith("http:"):
|
|
klass = linkcheck.checker.httpurl.HttpUrl
|
|
elif url.startswith("ftp:"):
|
|
klass = linkcheck.checker.ftpurl.FtpUrl
|
|
elif url.startswith("file:"):
|
|
klass = linkcheck.checker.fileurl.FileUrl
|
|
elif url.startswith("telnet:"):
|
|
klass = linkcheck.checker.telneturl.TelnetUrl
|
|
elif url.startswith("mailto:"):
|
|
klass = linkcheck.checker.mailtourl.MailtoUrl
|
|
elif url.startswith("gopher:"):
|
|
klass = linkcheck.checker.gopherurl.GopherUrl
|
|
elif url.startswith("https:"):
|
|
klass = linkcheck.checker.httpsurl.HttpsUrl
|
|
elif url.startswith("nntp:") or \
|
|
url.startswith("news:") or \
|
|
url.startswith("snews:"):
|
|
klass = linkcheck.checker.nntpurl.NntpUrl
|
|
elif ignored_schemes_re.search(url):
|
|
# ignored url
|
|
klass = linkcheck.checker.ignoredurl.IgnoredUrl
|
|
elif cmdline:
|
|
# assume local file on command line
|
|
klass = linkcheck.checker.fileurl.FileUrl
|
|
else:
|
|
# error url, no further checking, just log this
|
|
klass = linkcheck.checker.errorurl.ErrorUrl
|
|
url_data = klass(base_url, recursion_level, consumer,
|
|
parent_url=parent_url, base_ref=base_ref,
|
|
line=line, column=column, name=name)
|
|
if cmdline:
|
|
# add intern URL regex to config for every URL that was given
|
|
# on the command line
|
|
pat = url_data.get_intern_pattern()
|
|
linkcheck.log.debug(linkcheck.LOG_CMDLINE, "Pattern %r", pat)
|
|
if pat:
|
|
consumer.config_append('internlinks', linkcheck.get_link_pat(pat))
|
|
return url_data
|
|
|
|
|
|
def get_index_html (urls):
|
|
"""
|
|
Construct artificial index.html from given URLs.
|
|
|
|
@param urls: list with url strings
|
|
@type urls: list of string
|
|
"""
|
|
lines = ["<html>", "<body>"]
|
|
for entry in urls:
|
|
name = cgi.escape(entry)
|
|
url = cgi.escape(urllib.quote(entry))
|
|
lines.append('<a href="%s">%s</a>' % (url, name))
|
|
lines.extend(["</body>", "</html>"])
|
|
return os.linesep.join(lines)
|