| Left: | ||
| Right: |
| OLD | NEW |
|---|---|
| (Empty) | |
| 1 # coding: utf-8 | |
| 2 | |
| 3 # This Source Code is subject to the terms of the Mozilla Public License | |
| 4 # version 2.0 (the "License"). You can obtain a copy of the License at | |
| 5 # http://mozilla.org/MPL/2.0/. | |
| 6 | |
| 7 import MySQLdb, os, re, subprocess | |
| 8 from sitescripts.utils import get_config | |
| 9 | |
| 10 def hg(args): | |
| 11 return subprocess.Popen(["hg"] + args, stdout = subprocess.PIPE) | |
| 12 | |
| 13 def extract_urls(filter_list_dir): | |
| 14 os.chdir(filter_list_dir) | |
| 15 process = hg(["log", "--template", "{desc}\n"]) | |
| 16 urls = set([]) | |
| 17 | |
| 18 for line in process.stdout: | |
| 19 matches = re.match(r".*\b(https?://\S*)", line) | |
|
Wladimir Palant
2012/09/27 07:34:17
Please use re.search() rather than re.match() here
Felix Dahlke
2012/09/27 09:26:24
Done. You're right, I'm used to calling it "matche
| |
| 20 if not matches: | |
| 21 continue | |
| 22 | |
| 23 url = matches.group(1).strip() | |
| 24 urls.add(url) | |
| 25 | |
| 26 return urls | |
| 27 | |
| 28 def print_statements(urls): | |
| 29 for url in urls: | |
| 30 escaped_url = MySQLdb.escape_string(url) | |
| 31 print "INSERT INTO crawler_sites (url) VALUES ('" + escaped_url + "');" | |
| 32 | |
| 33 if __name__ == "__main__": | |
| 34 filter_list_dir = get_config().get("crawler", "filter_list_repository") | |
| 35 urls = extract_urls(filter_list_dir) | |
| 36 print_statements(urls) | |
| OLD | NEW |