mirror of
https://github.com/Hopiu/linkchecker.git
synced 2026-04-08 16:41:00 +00:00
anchor bug and test suite fixed
git-svn-id: https://linkchecker.svn.sourceforge.net/svnroot/linkchecker/trunk/linkchecker@236 e7d03fd6-7b0d-0410-9947-9c21f3af8025
This commit is contained in:
parent
e8e573ef4a
commit
8793b6cf01
13 changed files with 23 additions and 17 deletions
|
|
@ -1,3 +1,4 @@
|
|||
configure-stamp
|
||||
build-stamp
|
||||
.files-stamp
|
||||
*-out.*
|
||||
|
|
|
|||
7
Makefile
7
Makefile
|
|
@ -11,6 +11,8 @@ LCOPTS=-ocolored -Ftext -Fhtml -Fgml -Fsql -Fcsv -Fxml -R -t0 -v -s
|
|||
DEBPACKAGE = $(PACKAGE)_$(VERSION)_all.deb $(PACKAGE)-ssl_$(VERSION)_i386.deb
|
||||
PULLHOST=phoenix.net.uni-sb.de
|
||||
PULLPATH=/home/calvin/temp/linkchecker
|
||||
OFFLINETESTS = test_base test_misc test_file test_frames
|
||||
ONLINETESTS = test_mail test_http test_https test_news test_ftp
|
||||
|
||||
DESTDIR=/.
|
||||
.PHONY: test clean distclean package files upload dist locale all
|
||||
|
|
@ -69,7 +71,10 @@ uploadpull: distclean dist package files VERSION
|
|||
ssh -C -t shell1.sourceforge.net "cd /home/groups/$(PACKAGE) && make pull"
|
||||
|
||||
test:
|
||||
python2 test/regrtest.py
|
||||
python2 test/regrtest.py $(OFFLINE_TESTS)
|
||||
|
||||
onlinetest:
|
||||
python2 test/regrtest.py $(ONLINE_TESTS)
|
||||
|
||||
locale:
|
||||
$(MAKE) -C po
|
||||
|
|
|
|||
8
debian/changelog
vendored
8
debian/changelog
vendored
|
|
@ -4,12 +4,12 @@ linkchecker (1.3.0) unstable; urgency=low
|
|||
* fix agent matching in robotparser2.py
|
||||
* added more LinkPatterns (ripped from HTML::Tagset.pm)
|
||||
* fix ignored configfile settings for loggers
|
||||
* optional filename argument for -F (patch from
|
||||
* optional filename argument for -F (patch from
|
||||
Jamie Heilman <jamie@audible.transient.net>)
|
||||
* config file option to control which fields the loggers should print
|
||||
out
|
||||
* config file option to control which fields the loggers should print
|
||||
* fix anchor bug (unpack tuple of wrong size)
|
||||
|
||||
-- Bastian Kleineidam <calvin@users.sourceforge.net> Fri, 9 Feb 2001 10:51:24 +0100
|
||||
-- Bastian Kleineidam <calvin@users.sourceforge.net> Thu, 22 Feb 2001 01:35:50 +0100
|
||||
|
||||
linkchecker (1.2.14) unstable; urgency=low
|
||||
|
||||
|
|
|
|||
6
debian/control
vendored
6
debian/control
vendored
|
|
@ -2,13 +2,13 @@ Source: linkchecker
|
|||
Section: web
|
||||
Priority: optional
|
||||
Maintainer: Bastian Kleineidam <calvin@debian.org>
|
||||
Build-Depends: python2-base (>= 2.0), python2-dev (>= 2.0), debhelper (>= 3.0.0), libssl096-dev
|
||||
Build-Depends: python2-base, python2-dev, debhelper (>= 3.0.0), libssl096-dev
|
||||
Build-Depends-Indep: gettext
|
||||
Standards-Version: 3.5.1
|
||||
Standards-Version: 3.5.2
|
||||
|
||||
Package: linkchecker
|
||||
Architecture: all
|
||||
Depends: python2-base (>= 2.0)
|
||||
Depends: python2-base
|
||||
Suggests: linkchecker-ssl
|
||||
Description: check HTML documents for broken links
|
||||
Features:
|
||||
|
|
|
|||
|
|
@ -297,7 +297,7 @@ class UrlData:
|
|||
if not (anchor!="" and self.isHtml() and self.valid):
|
||||
return
|
||||
self.getContent()
|
||||
for cur_anchor,line in self.searchInForTag(AnchorPattern):
|
||||
for cur_anchor,line,name in self.searchInForTag(AnchorPattern):
|
||||
if cur_anchor == anchor:
|
||||
return
|
||||
self.setWarning("anchor #"+anchor+" not found")
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('file.html'):
|
||||
for file in ('file.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('frames.html'):
|
||||
for file in ('frames.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('ftp.html'):
|
||||
for file in ('ftp.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('http.html'):
|
||||
for file in ('http.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('https.html'):
|
||||
for file in ('https.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('mail.html'):
|
||||
for file in ('mail.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('misc.html'):
|
||||
for file in ('misc.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ config["anchors"] = 1
|
|||
config["verbose"] = 1
|
||||
config.disableThreading()
|
||||
htmldir = "test/html"
|
||||
for file in ('news.html'):
|
||||
for file in ('news.html',):
|
||||
url = os.path.join(htmldir, file)
|
||||
config.appendUrl(linkcheck.UrlData.GetUrlDataFrom(url, 0))
|
||||
linkcheck.checkUrls(config)
|
||||
|
|
|
|||
Loading…
Reference in a new issue