diff --git a/linkcheck/checker/ftpurl.py b/linkcheck/checker/ftpurl.py index 29f7cd13..2d6844a3 100644 --- a/linkcheck/checker/ftpurl.py +++ b/linkcheck/checker/ftpurl.py @@ -88,7 +88,6 @@ class FtpUrl(internpaturl.InternPatternUrl, proxysupport.ProxySupport): # note that the info may change every time a user logs in, # so don't add it to the url_data info. log.debug(LOG_CHECK, "FTP info %s", info) - pass else: raise LinkCheckerError(_("Got no answer from FTP server")) except EOFError as msg: @@ -103,7 +102,6 @@ class FtpUrl(internpaturl.InternPatternUrl, proxysupport.ProxySupport): features = self.url_connection.sendcmd("FEAT") except ftplib.error_perm as msg: log.debug(LOG_CHECK, "Ignoring error when getting FTP features: %s" % msg) - pass else: log.debug(LOG_CHECK, "FTP features %s", features) if " UTF-8" in features.splitlines(): diff --git a/linkcheck/checker/itmsservicesurl.py b/linkcheck/checker/itmsservicesurl.py index 982c8acb..deafd349 100644 --- a/linkcheck/checker/itmsservicesurl.py +++ b/linkcheck/checker/itmsservicesurl.py @@ -33,7 +33,6 @@ class ItmsServicesUrl(urlbase.UrlBase): def local_check(self): """Disable content checks.""" log.debug(LOG_CHECK, "Checking %s", self) - pass def check_content(self): """Allow recursion to check the url CGI param.""" diff --git a/linkcheck/checker/mailtourl.py b/linkcheck/checker/mailtourl.py index 512935b9..e29150e9 100644 --- a/linkcheck/checker/mailtourl.py +++ b/linkcheck/checker/mailtourl.py @@ -363,7 +363,6 @@ class MailtoUrl(urlbase.UrlBase): log.debug(LOG_CHECK, "found %d MX mailhosts:", len(answers)) for preference, host in mxdata: log.debug(LOG_CHECK, "MX host %r, preference %d", host, preference) - pass self.set_result(_("Valid mail address syntax")) def set_cache_url(self): diff --git a/linkcheck/configuration/__init__.py b/linkcheck/configuration/__init__.py index a4093b7b..a08f9d9a 100644 --- a/linkcheck/configuration/__init__.py +++ b/linkcheck/configuration/__init__.py @@ -470,7 +470,6 @@ def get_gnome_proxy(protocol="HTTP"): return "%s:%d" % (host, port) except Exception as msg: log.debug(LOG_CHECK, "error getting %s proxy from GNOME: %s", (protocol, msg)) - pass return None @@ -481,7 +480,6 @@ def get_kde_http_proxy(): return data.get("http_proxy") except Exception as msg: log.debug(LOG_CHECK, "error getting HTTP proxy from KDE: %s", msg) - pass def get_kde_ftp_proxy(): @@ -491,7 +489,6 @@ def get_kde_ftp_proxy(): return data.get("ftp_proxy") except Exception as msg: log.debug(LOG_CHECK, "error getting FTP proxy from KDE: %s", msg) - pass # The following KDE functions are largely ported and ajusted from diff --git a/linkcheck/robotparser2.py b/linkcheck/robotparser2.py index 1a5af387..1d827d76 100644 --- a/linkcheck/robotparser2.py +++ b/linkcheck/robotparser2.py @@ -197,7 +197,6 @@ class RobotFileParser: self.url, linenumber, ) - pass else: entry.rulelines.append(RuleLine(line[1], False)) state = 2 @@ -209,7 +208,6 @@ class RobotFileParser: self.url, linenumber, ) - pass else: entry.rulelines.append(RuleLine(line[1], True)) state = 2 @@ -221,7 +219,6 @@ class RobotFileParser: self.url, linenumber, ) - pass else: try: entry.crawldelay = max(0, int(line[1])) @@ -234,7 +231,6 @@ class RobotFileParser: linenumber, line[1], ) - pass elif line[0] == "sitemap": # Note that sitemap URLs must be absolute according to # http://www.sitemaps.org/protocol.html#submit_robots @@ -248,7 +244,6 @@ class RobotFileParser: linenumber, line[0], ) - pass else: log.debug( LOG_CHECK, @@ -257,7 +252,6 @@ class RobotFileParser: linenumber, line, ) - pass if state in (1, 2): self.entries.append(entry) self.modified()