From bc89449080ddacd1885e7ef7ac095d46a46077d7 Mon Sep 17 00:00:00 2001 From: calvin Date: Fri, 19 Nov 2004 11:06:31 +0000 Subject: [PATCH] add dots to info messages git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@2008 e7d03fd6-7b0d-0410-9947-9c21f3af8025 --- linkcheck/checker/httpurl.py | 18 +++++++++--------- linkcheck/checker/mailtourl.py | 2 +- linkcheck/checker/nntpurl.py | 4 ++-- linkcheck/checker/urlbase.py | 2 +- 4 files changed, 13 insertions(+), 13 deletions(-) diff --git a/linkcheck/checker/httpurl.py b/linkcheck/checker/httpurl.py index ea99c21f..1898b43f 100644 --- a/linkcheck/checker/httpurl.py +++ b/linkcheck/checker/httpurl.py @@ -117,18 +117,18 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): # set the proxy, so a 407 status after this is an error self.set_proxy(self.consumer.config["proxy"].get(self.scheme)) if self.proxy: - self.add_info(_("Using Proxy %r") % self.proxy) + self.add_info(_("Using Proxy %r.") % self.proxy) self.headers = None self.auth = None self.cookies = [] if not self.allows_robots(self.url): self.add_warning( - _("Access denied by robots.txt, checked only syntax")) + _("Access denied by robots.txt, checked only syntax.")) return if _is_amazon(self.urlparts[1]): self.add_warning(_("Amazon servers block HTTP HEAD requests, " - "using GET instead")) + "using GET instead.")) self.method = "GET" else: # first try with HEAD @@ -154,7 +154,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): if response.status == 305 and self.headers: oldproxy = (self.proxy, self.proxyauth) self.set_proxy(self.headers.getheader("Location")) - self.add_info(_("Enforced Proxy %r") % self.proxy) + self.add_info(_("Enforced Proxy %r.") % self.proxy) response = self._get_http_response() self.headers = response.msg self.proxy, self.proxyauth = oldproxy @@ -240,7 +240,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): newurl = unicode(newurl, "iso8859-1", "ignore") linkcheck.log.debug(linkcheck.LOG_CHECK, "Redirected to %r", newurl) - self.add_info(_("Redirected to %(url)s") % {'url': redirected}) + self.add_info(_("Redirected to %(url)s.") % {'url': redirected}) redirected, is_idn = linkcheck.url.url_norm(newurl) linkcheck.log.debug(linkcheck.LOG_CHECK, "Norm redirected to %r", redirected) @@ -248,12 +248,12 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): self.extern = self._get_extern(redirected) if self.is_extern(): self.add_info( - _("outside of domain filter, checked only syntax")) + _("Outside of domain filter, checked only syntax.")) return -1, response # check robots.txt allowance again if not self.allows_robots(redirected): self.add_warning( - _("Access denied by robots.txt, checked only syntax")) + _("Access denied by robots.txt, checked only syntax.")) return -1, response # see about recursive redirect all_seen = self.aliases + [self.cache_url_key] @@ -330,7 +330,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): # store cookies for valid links if self.consumer.config['cookies']: for c in self.cookies: - self.add_info("Cookie: %s" % c) + self.add_info("Cookie: %s." % c) try: out = self.consumer.cache.store_cookies(self.headers, self.urlparts[1]) @@ -345,7 +345,7 @@ class HttpUrl (urlbase.UrlBase, proxysupport.ProxySupport): self.set_result("OK") modified = self.headers.get('Last-Modified', '') if modified: - self.add_info(_("Last modified %s") % modified) + self.add_info(_("Last modified %s.") % modified) def _get_http_response (self): """Put request and return (status code, status text, mime object). diff --git a/linkcheck/checker/mailtourl.py b/linkcheck/checker/mailtourl.py index 9f189b00..89d729b0 100644 --- a/linkcheck/checker/mailtourl.py +++ b/linkcheck/checker/mailtourl.py @@ -129,7 +129,7 @@ class MailtoUrl (urlbase.UrlBase): linkcheck.log.debug(linkcheck.LOG_CHECK, "SMTP user info %r", info) if info[0] == 250: - self.add_info(_("Verified address: %(info)s") % \ + self.add_info(_("Verified address: %(info)s.") % \ {'info': str(info[1])}) except smtplib.SMTPException, msg: self.add_warning( diff --git a/linkcheck/checker/nntpurl.py b/linkcheck/checker/nntpurl.py index 63682650..d75fdd62 100644 --- a/linkcheck/checker/nntpurl.py +++ b/linkcheck/checker/nntpurl.py @@ -109,14 +109,14 @@ class NntpUrl (urlbase.UrlBase): if '@' in group: # request article resp, number, mid = nntp.stat("<"+group+">") - self.add_info(_('Articel number %s found') % number) + self.add_info(_('Articel number %s found.') % number) else: # split off trailing articel span group = group.split('/', 1)[0] if group: # request group info resp, count, first, last, name = nntp.group(group) - self.add_info(_("Group %s has %s articles, range %s to %s") %\ + self.add_info(_("Group %s has %s articles, range %s to %s.")%\ (name, count, first, last)) else: # group name is the empty string diff --git a/linkcheck/checker/urlbase.py b/linkcheck/checker/urlbase.py index ca436d1a..31a7b969 100644 --- a/linkcheck/checker/urlbase.py +++ b/linkcheck/checker/urlbase.py @@ -323,7 +323,7 @@ class UrlBase (object): time.sleep(self.consumer.config['wait']) t = time.time() if self.is_extern(): - self.add_info(_("outside of domain filter, checked only syntax")) + self.add_info(_("Outside of domain filter, checked only syntax.")) return # check connection