diff --git a/reppy/rep-cpp b/reppy/rep-cpp index 2439eb6..e13fbe2 160000 --- a/reppy/rep-cpp +++ b/reppy/rep-cpp @@ -1 +1 @@ -Subproject commit 2439eb64a3fba37981be6f9e2a48934f2176bb82 +Subproject commit e13fbe2e2844d7d5ad4e01087f9b4b888589a1e8 diff --git a/setup.py b/setup.py index 62a9922..237be36 100644 --- a/setup.py +++ b/setup.py @@ -54,7 +54,7 @@ setup( name = 'reppy', - version = '0.4.2', + version = '0.4.3', description = 'Replacement robots.txt Parser', long_description = '''Replaces the built-in robotsparser with a RFC-conformant implementation that supports modern robots.txt constructs like diff --git a/tests/test_robots.py b/tests/test_robots.py index e76e438..fd05433 100644 --- a/tests/test_robots.py +++ b/tests/test_robots.py @@ -136,7 +136,7 @@ def test_exposes_sitemaps(self): def test_case_insensitivity(self): '''Make sure user agent matches are case insensitive''' robot = robots.Robots.parse('http://example.com/robots.txt', ''' - User-agent: agent + User-agent: Agent Disallow: /path ''') self.assertFalse(robot.allowed('/path', 'agent'))