summaryrefslogtreecommitdiff
path: root/www
diff options
context:
space:
mode:
authorMikhail Teterin <mi@FreeBSD.org>2001-07-19 16:34:13 +0000
committerMikhail Teterin <mi@FreeBSD.org>2001-07-19 16:34:13 +0000
commitc63c60cc72f112cca580c3c24febf500a6058992 (patch)
tree2a16dfc5b8001cf547c708e2a30ea240a1b8fd6d /www
parentAdd the ipv6 category. (diff)
Add another patch (by Mike Meyer), who answered my call for a
Python guru and fixed the linbot to work under modern version of Python (as well as the 1.5). He also replaced regex and regsub with the re module, which seems to make it work better with no-ascii web-pages too (a nice side-effect). Bump up PORTREVISION and give MAINTAINERship to Mike (he *did* see it coming). Obtained from: Mike Meyer
Notes
Notes: svn path=/head/; revision=45250
Diffstat (limited to 'www')
-rw-r--r--www/linbot/Makefile6
-rw-r--r--www/linbot/files/patch-compat90
-rw-r--r--www/webcheck/Makefile6
-rw-r--r--www/webcheck/files/patch-compat90
4 files changed, 186 insertions, 6 deletions
diff --git a/www/linbot/Makefile b/www/linbot/Makefile
index e41899f6f48f..c39f4b307ea4 100644
--- a/www/linbot/Makefile
+++ b/www/linbot/Makefile
@@ -7,15 +7,15 @@
PORTNAME= linbot
PORTVERSION= 1.0
-PORTREVISION= 2
+PORTREVISION= 3
CATEGORIES= www python
MASTER_SITES= ${MASTER_SITE_SUNSITE}
MASTER_SITE_SUBDIR= apps/www/misc
EXTRACT_SUFX= .tgz
-MAINTAINER= ports@FreeBSD.org
+MAINTAINER= mwm@mired.org
-RUN_DEPENDS= python:${PORTSDIR}/lang/python15
+USE_PYTHON= YES
do-install:
${MKDIR} ${PREFIX}/share/linbot
diff --git a/www/linbot/files/patch-compat b/www/linbot/files/patch-compat
new file mode 100644
index 000000000000..81dc79d37b38
--- /dev/null
+++ b/www/linbot/files/patch-compat
@@ -0,0 +1,90 @@
+diff -ru /tmp/lb/linbot-1.0/myUrlLib.py ./myUrlLib.py
+--- myUrlLib.py Sun Mar 14 13:24:20 1999
++++ myUrlLib.py Wed Jul 18 20:42:57 2001
+@@ -37,6 +37,7 @@
+ import htmlparse
+ import debugio
+ import sys
++import socket
+
+
+ def get_robots(location):
+@@ -105,7 +106,10 @@
+
+ if (parent is None):
+ Link.baseurl=self.URL
+- Link.base=self.URL[:string.rfind(self.URL,'/')+1]
++ if hasattr(self.URL, 'rfind'):
++ Link.base=self.URL[:self.URL.rfind('/')+1]
++ else:
++ Link.base=self.URL[:string.rfind(self.URL,'/')+1]
+ if Link.base[-2:] == '//': Link.base = self.URL
+ debugio.write('\tbase: %s' % Link.base)
+ if self.scheme == 'http':
+--- robotparser.py Sat Jan 9 19:01:45 1999
++++ robotparser.py Wed Jul 18 20:29:13 2001
+@@ -37,7 +37,7 @@
+ self.parse(urlopener.open(self.url).readlines())
+
+ def parse(self, lines):
+- import regsub, string, regex
++ import re, string
+ active = []
+ for line in lines:
+ if self.debug: print '>', line,
+@@ -49,7 +49,7 @@
+ line = string.strip(line[:string.find(line, '#')])
+ if not line:
+ continue
+- line = regsub.split(line, ' *: *')
++ line = re.split(' *: *', line)
+ if len(line) == 2:
+ line[0] = string.lower(line[0])
+ if line[0] == 'user-agent':
+@@ -62,7 +62,7 @@
+ if line[1]:
+ if self.debug: print '>> disallow:', line[1]
+ for agent in active:
+- self.rules[agent].append(regex.compile(line[1]))
++ self.rules[agent].append(re.compile(line[1]))
+ else:
+ pass
+ for agent in active:
+@@ -83,7 +83,7 @@
+ return 1
+ path = urlparse.urlparse(url)[2]
+ for rule in self.rules[ag]:
+- if rule.match(path) != -1:
++ if rule.match(path):
+ if self.debug: print '>> disallowing', url, 'fetch by', agent
+ return 0
+ if self.debug: print '>> allowing', url, 'fetch by', agent
+--- schemes/filelink.py Thu Mar 11 22:56:07 1999
++++ schemes/filelink.py Wed Jul 18 20:19:26 2001
+@@ -42,7 +42,7 @@
+ import time
+ import mimetypes
+ import myUrlLib
+-import regsub
++import re
+
+ mimetypes.types_map['.shtml']='text/html'
+
+@@ -51,7 +51,7 @@
+ parsed = urlparse.urlparse(self.URL,'file',0)
+ filename = parsed[2]
+ if os.name != 'posix':
+- filename = regsub.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
++ filename = re.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
+ try:
+ stats = os.stat(filename)
+ except os.error:
+@@ -70,7 +70,7 @@
+ parsed = urlparse.urlparse(url,'file',0)
+ filename = parsed[2]
+ if os.name != 'posix':
+- filename = regsub.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
++ filename = re.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
+
+ return open(filename,'r').read()
+
diff --git a/www/webcheck/Makefile b/www/webcheck/Makefile
index e41899f6f48f..c39f4b307ea4 100644
--- a/www/webcheck/Makefile
+++ b/www/webcheck/Makefile
@@ -7,15 +7,15 @@
PORTNAME= linbot
PORTVERSION= 1.0
-PORTREVISION= 2
+PORTREVISION= 3
CATEGORIES= www python
MASTER_SITES= ${MASTER_SITE_SUNSITE}
MASTER_SITE_SUBDIR= apps/www/misc
EXTRACT_SUFX= .tgz
-MAINTAINER= ports@FreeBSD.org
+MAINTAINER= mwm@mired.org
-RUN_DEPENDS= python:${PORTSDIR}/lang/python15
+USE_PYTHON= YES
do-install:
${MKDIR} ${PREFIX}/share/linbot
diff --git a/www/webcheck/files/patch-compat b/www/webcheck/files/patch-compat
new file mode 100644
index 000000000000..81dc79d37b38
--- /dev/null
+++ b/www/webcheck/files/patch-compat
@@ -0,0 +1,90 @@
+diff -ru /tmp/lb/linbot-1.0/myUrlLib.py ./myUrlLib.py
+--- myUrlLib.py Sun Mar 14 13:24:20 1999
++++ myUrlLib.py Wed Jul 18 20:42:57 2001
+@@ -37,6 +37,7 @@
+ import htmlparse
+ import debugio
+ import sys
++import socket
+
+
+ def get_robots(location):
+@@ -105,7 +106,10 @@
+
+ if (parent is None):
+ Link.baseurl=self.URL
+- Link.base=self.URL[:string.rfind(self.URL,'/')+1]
++ if hasattr(self.URL, 'rfind'):
++ Link.base=self.URL[:self.URL.rfind('/')+1]
++ else:
++ Link.base=self.URL[:string.rfind(self.URL,'/')+1]
+ if Link.base[-2:] == '//': Link.base = self.URL
+ debugio.write('\tbase: %s' % Link.base)
+ if self.scheme == 'http':
+--- robotparser.py Sat Jan 9 19:01:45 1999
++++ robotparser.py Wed Jul 18 20:29:13 2001
+@@ -37,7 +37,7 @@
+ self.parse(urlopener.open(self.url).readlines())
+
+ def parse(self, lines):
+- import regsub, string, regex
++ import re, string
+ active = []
+ for line in lines:
+ if self.debug: print '>', line,
+@@ -49,7 +49,7 @@
+ line = string.strip(line[:string.find(line, '#')])
+ if not line:
+ continue
+- line = regsub.split(line, ' *: *')
++ line = re.split(' *: *', line)
+ if len(line) == 2:
+ line[0] = string.lower(line[0])
+ if line[0] == 'user-agent':
+@@ -62,7 +62,7 @@
+ if line[1]:
+ if self.debug: print '>> disallow:', line[1]
+ for agent in active:
+- self.rules[agent].append(regex.compile(line[1]))
++ self.rules[agent].append(re.compile(line[1]))
+ else:
+ pass
+ for agent in active:
+@@ -83,7 +83,7 @@
+ return 1
+ path = urlparse.urlparse(url)[2]
+ for rule in self.rules[ag]:
+- if rule.match(path) != -1:
++ if rule.match(path):
+ if self.debug: print '>> disallowing', url, 'fetch by', agent
+ return 0
+ if self.debug: print '>> allowing', url, 'fetch by', agent
+--- schemes/filelink.py Thu Mar 11 22:56:07 1999
++++ schemes/filelink.py Wed Jul 18 20:19:26 2001
+@@ -42,7 +42,7 @@
+ import time
+ import mimetypes
+ import myUrlLib
+-import regsub
++import re
+
+ mimetypes.types_map['.shtml']='text/html'
+
+@@ -51,7 +51,7 @@
+ parsed = urlparse.urlparse(self.URL,'file',0)
+ filename = parsed[2]
+ if os.name != 'posix':
+- filename = regsub.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
++ filename = re.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
+ try:
+ stats = os.stat(filename)
+ except os.error:
+@@ -70,7 +70,7 @@
+ parsed = urlparse.urlparse(url,'file',0)
+ filename = parsed[2]
+ if os.name != 'posix':
+- filename = regsub.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
++ filename = re.sub("^/\(//\)?\([a-zA-Z]\)[|:]","\\2:",filename)
+
+ return open(filename,'r').read()
+