| Left: | ||
| Right: |
| OLD | NEW |
|---|---|
| 1 # This file is part of the Adblock Plus web scripts, | 1 # This file is part of the Adblock Plus web scripts, |
| 2 # Copyright (C) 2006-present eyeo GmbH | 2 # Copyright (C) 2006-present eyeo GmbH |
| 3 # | 3 # |
| 4 # Adblock Plus is free software: you can redistribute it and/or modify | 4 # Adblock Plus is free software: you can redistribute it and/or modify |
| 5 # it under the terms of the GNU General Public License version 3 as | 5 # it under the terms of the GNU General Public License version 3 as |
| 6 # published by the Free Software Foundation. | 6 # published by the Free Software Foundation. |
| 7 # | 7 # |
| 8 # Adblock Plus is distributed in the hope that it will be useful, | 8 # Adblock Plus is distributed in the hope that it will be useful, |
| 9 # but WITHOUT ANY WARRANTY; without even the implied warranty of | 9 # but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 10 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | 10 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 11 # GNU General Public License for more details. | 11 # GNU General Public License for more details. |
| 12 # | 12 # |
| 13 # You should have received a copy of the GNU General Public License | 13 # You should have received a copy of the GNU General Public License |
| 14 # along with Adblock Plus. If not, see <http://www.gnu.org/licenses/>. | 14 # along with Adblock Plus. If not, see <http://www.gnu.org/licenses/>. |
| 15 | 15 |
| 16 import os | 16 import os |
| 17 import re | 17 import re |
| 18 import errno | 18 import errno |
| 19 import codecs | 19 import codecs |
| 20 import ConfigParser | 20 import ConfigParser |
| 21 import logging | 21 import logging |
| 22 import lxml.html | |
| 23 | |
| 22 from argparse import ArgumentParser | 24 from argparse import ArgumentParser |
| 23 | 25 |
| 24 from cms.utils import get_page_params, process_page | 26 from cms.utils import get_page_params, process_page |
| 25 from cms.sources import create_source | 27 from cms.sources import create_source |
| 26 | 28 |
| 27 MIN_TRANSLATED = 0.3 | 29 MIN_TRANSLATED = 0.3 |
| 28 | 30 |
| 29 | 31 |
| 30 def generate_pages(repo, output_dir, revision): | 32 def generate_pages(repo, output_dir, revision, relative): |
| 31 known_files = set() | 33 known_files = set() |
| 32 | 34 |
| 33 def write_file(path_parts, contents, binary=False): | 35 def write_file(path_parts, contents, binary=False): |
| 34 encoding = None if binary else 'utf-8' | 36 encoding = None if binary else 'utf-8' |
| 35 outfile = os.path.join(output_dir, *path_parts) | 37 outfile = os.path.join(output_dir, *path_parts) |
| 36 if outfile in known_files: | 38 if outfile in known_files: |
| 37 logging.warning('File %s has multiple sources', outfile) | 39 logging.warning('File %s has multiple sources', outfile) |
| 38 return | 40 return |
| 39 known_files.add(outfile) | 41 known_files.add(outfile) |
| 40 | 42 |
| (...skipping 42 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 83 orig_has_locale = source.has_locale | 85 orig_has_locale = source.has_locale |
| 84 | 86 |
| 85 def has_locale(locale, page): | 87 def has_locale(locale, page): |
| 86 page = get_locale_file(page) | 88 page = get_locale_file(page) |
| 87 if (locale, page) in blacklist: | 89 if (locale, page) in blacklist: |
| 88 return False | 90 return False |
| 89 return orig_has_locale(locale, page) | 91 return orig_has_locale(locale, page) |
| 90 source.has_locale = has_locale | 92 source.has_locale = has_locale |
| 91 source.resolve_link.cache_clear() | 93 source.resolve_link.cache_clear() |
| 92 | 94 |
| 95 def rewrite_link(link): | |
|
Vasily Kuznetsov
2017/11/14 17:30:27
This seems like it would work but the links would
| |
| 96 if link.startswith('/'): | |
| 97 if source.version: | |
| 98 link += '?' + source.version | |
| 99 if relative: | |
| 100 depth = len(page.split('/')) | |
| 101 link = '/'.join(['..'] * depth) + link | |
| 102 return link | |
| 103 | |
| 93 # Second pass: actually generate pages this time | 104 # Second pass: actually generate pages this time |
| 94 for locale, page in pagelist: | 105 for locale, page in pagelist: |
| 95 pagedata = process_page(source, locale, page) | 106 pagedata = process_page(source, locale, page) |
| 96 | 107 |
| 97 # Make sure links to static files are versioned | 108 root = lxml.html.fromstring(pagedata).getroottree() |
| 98 pagedata = re.sub(r'(<script\s[^<>]*\bsrc="/[^"<>]+)', r'\1?%s' % so urce.version, pagedata) | |
| 99 pagedata = re.sub(r'(<link\s[^<>]*\bhref="/[^"<>]+)', r'\1?%s' % sou rce.version, pagedata) | |
| 100 pagedata = re.sub(r'(<img\s[^<>]*\bsrc="/[^"<>]+)', r'\1?%s' % sourc e.version, pagedata) | |
| 101 | 109 |
| 110 expr = '//img[@src|@srcset]|//script[@src]|//link[@href]|//a[@href]' | |
| 111 for elem in root.xpath(expr): | |
| 112 if 'srcset' in elem.attrib: | |
| 113 srcset = [] | |
| 114 for src_zoom in elem.attrib['srcset'].split(', '): | |
|
Vasily Kuznetsov
2017/11/14 17:30:28
Is the space after the comma mandatory? Otherwise
| |
| 115 try: | |
| 116 src, zoom = src_zoom.split(' ') | |
| 117 except ValueError: | |
| 118 src = src_zoom | |
| 119 zoom = None | |
| 120 res = rewrite_link(src) | |
| 121 if zoom: | |
| 122 res = res + ' ' + zoom | |
| 123 srcset.append(res) | |
| 124 | |
| 125 elem.attrib['srcset'] = ', '.join(srcset) | |
| 126 | |
| 127 for key in ['src', 'href']: | |
| 128 if key in elem.attrib: | |
| 129 elem.attrib[key] = rewrite_link(elem.attrib[key]) | |
| 130 | |
| 131 pagedata = lxml.html.tostring(root, encoding='unicode') | |
| 102 write_file([locale] + page.split('/'), pagedata) | 132 write_file([locale] + page.split('/'), pagedata) |
| 103 | 133 |
| 104 for filename in source.list_localizable_files(): | 134 for filename in source.list_localizable_files(): |
| 105 for locale in locales: | 135 for locale in locales: |
| 106 if source.has_localizable_file(locale, filename): | 136 if source.has_localizable_file(locale, filename): |
| 107 filedata = source.read_localizable_file(locale, filename) | 137 filedata = source.read_localizable_file(locale, filename) |
| 108 write_file([locale] + filename.split('/'), filedata, binary= True) | 138 write_file([locale] + filename.split('/'), filedata, binary= True) |
| 109 | 139 |
| 110 for filename in source.list_static(): | 140 for filename in source.list_static(): |
| 111 write_file(filename.split('/'), source.read_static(filename), binary =True) | 141 write_file(filename.split('/'), source.read_static(filename), binary =True) |
| (...skipping 10 matching lines...) Expand all Loading... | |
| 122 os.rmdir(path) | 152 os.rmdir(path) |
| 123 remove_unknown(output_dir) | 153 remove_unknown(output_dir) |
| 124 | 154 |
| 125 | 155 |
| 126 if __name__ == '__main__': | 156 if __name__ == '__main__': |
| 127 parser = ArgumentParser('Convert website source to static website') | 157 parser = ArgumentParser('Convert website source to static website') |
| 128 parser.add_argument('-r', '--rev', | 158 parser.add_argument('-r', '--rev', |
| 129 help=('Specify which revision to generate from. ' | 159 help=('Specify which revision to generate from. ' |
| 130 'See "hg help revisions" for details.'), | 160 'See "hg help revisions" for details.'), |
| 131 default='default') | 161 default='default') |
| 162 parser.add_argument('--relative', action='store_true', default=False) | |
| 132 parser.add_argument('source', help="Path to website's repository") | 163 parser.add_argument('source', help="Path to website's repository") |
| 133 parser.add_argument('output', help='Path to desired output directory') | 164 parser.add_argument('output', help='Path to desired output directory') |
| 134 args = parser.parse_args() | 165 args = parser.parse_args() |
| 135 generate_pages(args.source, args.output, args.rev) | 166 generate_pages(args.source, args.output, args.rev, args.relative) |
| OLD | NEW |