mirror of
https://github.com/Hopiu/linkchecker.git
synced 2026-04-09 17:10:58 +00:00
add dots to info messages
git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@2008 e7d03fd6-7b0d-0410-9947-9c21f3af8025
This commit is contained in:
parent
edc8b4f48e
commit
bc89449080
4 changed files with 13 additions and 13 deletions
|
|
@ -117,18 +117,18 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
# set the proxy, so a 407 status after this is an error
|
||||
self.set_proxy(self.consumer.config["proxy"].get(self.scheme))
|
||||
if self.proxy:
|
||||
self.add_info(_("Using Proxy %r") % self.proxy)
|
||||
self.add_info(_("Using Proxy %r.") % self.proxy)
|
||||
self.headers = None
|
||||
self.auth = None
|
||||
self.cookies = []
|
||||
if not self.allows_robots(self.url):
|
||||
self.add_warning(
|
||||
_("Access denied by robots.txt, checked only syntax"))
|
||||
_("Access denied by robots.txt, checked only syntax."))
|
||||
return
|
||||
|
||||
if _is_amazon(self.urlparts[1]):
|
||||
self.add_warning(_("Amazon servers block HTTP HEAD requests, "
|
||||
"using GET instead"))
|
||||
"using GET instead."))
|
||||
self.method = "GET"
|
||||
else:
|
||||
# first try with HEAD
|
||||
|
|
@ -154,7 +154,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
if response.status == 305 and self.headers:
|
||||
oldproxy = (self.proxy, self.proxyauth)
|
||||
self.set_proxy(self.headers.getheader("Location"))
|
||||
self.add_info(_("Enforced Proxy %r") % self.proxy)
|
||||
self.add_info(_("Enforced Proxy %r.") % self.proxy)
|
||||
response = self._get_http_response()
|
||||
self.headers = response.msg
|
||||
self.proxy, self.proxyauth = oldproxy
|
||||
|
|
@ -240,7 +240,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
newurl = unicode(newurl, "iso8859-1", "ignore")
|
||||
linkcheck.log.debug(linkcheck.LOG_CHECK, "Redirected to %r",
|
||||
newurl)
|
||||
self.add_info(_("Redirected to %(url)s") % {'url': redirected})
|
||||
self.add_info(_("Redirected to %(url)s.") % {'url': redirected})
|
||||
redirected, is_idn = linkcheck.url.url_norm(newurl)
|
||||
linkcheck.log.debug(linkcheck.LOG_CHECK, "Norm redirected to %r",
|
||||
redirected)
|
||||
|
|
@ -248,12 +248,12 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
self.extern = self._get_extern(redirected)
|
||||
if self.is_extern():
|
||||
self.add_info(
|
||||
_("outside of domain filter, checked only syntax"))
|
||||
_("Outside of domain filter, checked only syntax."))
|
||||
return -1, response
|
||||
# check robots.txt allowance again
|
||||
if not self.allows_robots(redirected):
|
||||
self.add_warning(
|
||||
_("Access denied by robots.txt, checked only syntax"))
|
||||
_("Access denied by robots.txt, checked only syntax."))
|
||||
return -1, response
|
||||
# see about recursive redirect
|
||||
all_seen = self.aliases + [self.cache_url_key]
|
||||
|
|
@ -330,7 +330,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
# store cookies for valid links
|
||||
if self.consumer.config['cookies']:
|
||||
for c in self.cookies:
|
||||
self.add_info("Cookie: %s" % c)
|
||||
self.add_info("Cookie: %s." % c)
|
||||
try:
|
||||
out = self.consumer.cache.store_cookies(self.headers,
|
||||
self.urlparts[1])
|
||||
|
|
@ -345,7 +345,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport):
|
|||
self.set_result("OK")
|
||||
modified = self.headers.get('Last-Modified', '')
|
||||
if modified:
|
||||
self.add_info(_("Last modified %s") % modified)
|
||||
self.add_info(_("Last modified %s.") % modified)
|
||||
|
||||
def _get_http_response (self):
|
||||
"""Put request and return (status code, status text, mime object).
|
||||
|
|
|
|||
|
|
@ -129,7 +129,7 @@ class MailtoUrl (urlbase.UrlBase):
|
|||
linkcheck.log.debug(linkcheck.LOG_CHECK,
|
||||
"SMTP user info %r", info)
|
||||
if info[0] == 250:
|
||||
self.add_info(_("Verified address: %(info)s") % \
|
||||
self.add_info(_("Verified address: %(info)s.") % \
|
||||
{'info': str(info[1])})
|
||||
except smtplib.SMTPException, msg:
|
||||
self.add_warning(
|
||||
|
|
|
|||
|
|
@ -109,14 +109,14 @@ class NntpUrl (urlbase.UrlBase):
|
|||
if '@' in group:
|
||||
# request article
|
||||
resp, number, mid = nntp.stat("<"+group+">")
|
||||
self.add_info(_('Articel number %s found') % number)
|
||||
self.add_info(_('Articel number %s found.') % number)
|
||||
else:
|
||||
# split off trailing articel span
|
||||
group = group.split('/', 1)[0]
|
||||
if group:
|
||||
# request group info
|
||||
resp, count, first, last, name = nntp.group(group)
|
||||
self.add_info(_("Group %s has %s articles, range %s to %s") %\
|
||||
self.add_info(_("Group %s has %s articles, range %s to %s.")%\
|
||||
(name, count, first, last))
|
||||
else:
|
||||
# group name is the empty string
|
||||
|
|
|
|||
|
|
@ -323,7 +323,7 @@ class UrlBase (object):
|
|||
time.sleep(self.consumer.config['wait'])
|
||||
t = time.time()
|
||||
if self.is_extern():
|
||||
self.add_info(_("outside of domain filter, checked only syntax"))
|
||||
self.add_info(_("Outside of domain filter, checked only syntax."))
|
||||
return
|
||||
|
||||
# check connection
|
||||
|
|
|
|||
Loading…
Reference in a new issue