linkchecker/linkcheck/checker/ftpurl.py
calvin 3eac1be9ab Require and use Python 2.5
Use Python 2.5 features and get rid of old compat code. Also some
code cleanups have been made.


git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@3737 e7d03fd6-7b0d-0410-9947-9c21f3af8025
2008-04-27 11:39:21 +00:00

256 lines
8.2 KiB
Python

# -*- coding: iso-8859-1 -*-
# Copyright (C) 2000-2008 Bastian Kleineidam
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
"""
Handle FTP links.
"""
import ftplib
import time
import urllib
import cStringIO as StringIO
from .. import log, LOG_CHECK
import linkcheck
import proxysupport
import httpurl
import internpaturl
import linkcheck.ftpparse._ftpparse as ftpparse
from const import WARN_FTP_MISSING_SLASH, PARSE_EXTENSIONS
DEFAULT_TIMEOUT_SECS = 300
class FtpUrl (internpaturl.InternPatternUrl, proxysupport.ProxySupport):
"""
Url link with ftp scheme.
"""
def reset (self):
"""
Initialize FTP url data.
"""
super(FtpUrl, self).reset()
# list of files for recursion
self.files = []
# last part of URL filename
self.filename = None
def check_connection (self):
"""
In case of proxy, delegate to HttpUrl. Else check in this
order: login, changing directory, list the file.
"""
# proxy support (we support only http)
self.set_proxy(self.aggregate.config["proxy"].get(self.scheme))
if self.proxy:
# using a (HTTP) proxy
http = httpurl.HttpUrl(self.base_url,
self.recursion_level,
self.aggregate,
parent_url=self.parent_url,
base_ref=self.base_ref,
line=self.line,
column=self.column,
name=self.name)
http.build_url()
return http.check()
self.login()
self.filename = self.cwd()
self.listfile()
self.files = []
return None
def get_user_password (self):
"""
Get credentials to use for login.
"""
if self.userinfo:
return urllib.splitpasswd(self.userinfo)
return super(FtpUrl, self).get_user_password()
def login (self):
"""
Log into ftp server and check the welcome message.
"""
# ready to connect
_user, _password = self.get_user_password()
host = self.urlparts[1]
key = ("ftp", host, _user, _password)
conn = self.aggregate.connections.get(key)
if conn is not None and conn.sock is not None:
# reuse cached FTP connection
self.url_connection = conn
return
self.aggregate.connections.wait_for_host(host)
try:
self.url_connection = ftplib.FTP()
if log.is_debug(LOG_CHECK):
self.url_connection.set_debuglevel(1)
self.url_connection.connect(host)
if _user is None:
self.url_connection.login()
elif _password is None:
self.url_connection.login(_user)
else:
self.url_connection.login(_user, _password)
except EOFError, msg:
msg = str(msg)
raise linkcheck.LinkCheckerError(
_("Remote host has closed connection: %r") % msg)
if not self.url_connection.getwelcome():
raise linkcheck.LinkCheckerError(
_("Got no answer from FTP server"))
# don't set info anymore, this may change every time we log in
#self.add_info(linkcheck.strformat.unicode_safe(info))
def cwd (self):
"""
Change to URL parent directory. Return filename of last path
component.
"""
dirname = self.urlparts[2].strip('/')
dirs = dirname.split('/')
filename = dirs.pop()
self.url_connection.cwd('/')
for d in dirs:
self.url_connection.cwd(d)
return filename
def listfile (self):
"""
See if filename is in the current FTP directory.
"""
if not self.filename:
return
files = self.get_files()
log.debug(LOG_CHECK, "FTP files %s", str(files))
if self.filename in files:
# file found
return
# it could be a directory if the trailing slash was forgotten
if "%s/" % self.filename in files:
if not self.url.endswith('/'):
self.add_warning(
_("Missing trailing directory slash in ftp url."),
tag=WARN_FTP_MISSING_SLASH)
self.url += '/'
return
raise ftplib.error_perm("550 File not found")
def get_files (self):
"""
Get list of filenames in directory. Subdirectories have an
ending slash.
"""
files = []
def add_entry (line):
"""
Parse list line and add the entry it points to to the file list.
"""
log.debug(LOG_CHECK, "Directory entry %r", line)
try:
fpo = ftpparse.parse(line)
name = fpo.name
if fpo.trycwd:
name += "/"
if fpo.trycwd or fpo.tryretr:
files.append(name)
except (ValueError, AttributeError), msg:
log.debug(LOG_CHECK, "%s (%s)", str(msg), line)
self.url_connection.dir(add_entry)
return files
def is_html (self):
"""
See if URL target is a HTML file by looking at the extension.
"""
return bool(PARSE_EXTENSIONS['html'].search(self.url))
def is_css (self):
"""
See if URL target is a CSS file by looking at the extension.
"""
return bool(PARSE_EXTENSIONS['css'].search(self.url))
def is_parseable (self):
"""
See if URL target is parseable for recursion.
"""
if self.is_directory():
return True
for ro in PARSE_EXTENSIONS.values():
if ro.search(self.url):
return True
return False
def is_directory (self):
"""
See if URL target is a directory.
"""
return self.url.endswith('/')
def parse_url (self):
"""
Parse URL target for links.
"""
if self.is_directory():
self.parse_html()
return
for key, ro in PARSE_EXTENSIONS.items():
if ro.search(self.url):
getattr(self, "parse_"+key)()
def get_content (self):
"""
Return URL target content, or in case of directories a dummy HTML
file with links to the files.
"""
if not self.valid:
return ""
if self.data is not None:
return self.data
t = time.time()
if self.is_directory():
self.url_connection.cwd(self.filename)
self.files = self.get_files()
self.data = linkcheck.checker.get_index_html(self.files)
else:
# download file in BINARY mode
ftpcmd = "RETR %s" % self.filename
buf = StringIO.StringIO()
def stor_data (s):
"""Helper method storing given data"""
buf.write(s)
self.url_connection.retrbinary(ftpcmd, stor_data)
self.data = buf.getvalue()
buf.close()
self.dltime = time.time() - t
self.dlsize = len(self.data)
return self.data
def close_connection (self):
"""
Add the open connection to the connection pool.
"""
if self.url_connection is None:
return
# add to cached connections
_user, _password = self.get_user_password()
key = ("ftp", self.urlparts[1], _user, _password)
cache_add = self.aggregate.connections.add
cache_add(key, self.url_connection, DEFAULT_TIMEOUT_SECS)
self.url_connection = None