1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556 |
- #!/usr/bin/env python
- # Copyright (c) 2011 The Chromium Authors. All rights reserved.
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
- import os
- import tempfile
- import unittest
- import webforms_aggregator
- class WebformsAggregatorTest(unittest.TestCase):
- """Unit tests for the webforms_aggregator module."""
- def setUp(self):
- self.cookie_file = 'test.cookie'
- self.url1 = 'http://www.google.com'
- self.url2 = 'http://www.macys.com'
- self.domain = 'google.com'
- self.url_file = tempfile.NamedTemporaryFile(suffix='.txt', delete=False)
- self.url_file.file.write(
- 'URLs to crawl:\n%s\n%s\n' % (self.url1, self.url2))
- self.url_file.close()
- def tearDown(self):
- if os.path.isfile(self.cookie_file):
- os.unlink(self.cookie_file)
- if os.path.isfile(self.url_file.name):
- self.url_file.close()
- os.unlink(self.url_file.name)
- def testRetrieverDownloadsPage(self):
- """Verify the retriever can download a page."""
- r = webforms_aggregator.Retriever(self.url1, self.domain, self.cookie_file)
- self.assertTrue(r.Download(),
- msg='Retriever could not download "%s"' % self.url1)
- def testCrawlerFindsRegPageFromUrl(self):
- """Verify that the crawler is able to find a reg page from the given URL."""
- c = webforms_aggregator.Crawler(self.url2)
- self.assertTrue(
- c.Run(), msg='Crawler could not find the reg page of "%s"' % self.url2)
- def testThreadedCrawlerFindsRegPageFromUrlsFile(self):
- """Verify the threaded crawler finds reg page from a file of URLs."""
- c = webforms_aggregator.ThreadedCrawler(self.url_file.name)
- self.assertNotEqual(
- c.Run(), -1,
- msg='Threaded crawler could not find the reg page from the URLs file')
- if __name__ == '__main__':
- suite = unittest.TestLoader().loadTestsFromTestCase(
- WebformsAggregatorTest)
- unittest.TextTestRunner(verbosity=2).run(suite)
|