| Left: | ||
| Right: |
| OLD | NEW |
|---|---|
| 1 # This Source Code Form is subject to the terms of the Mozilla Public | 1 # This Source Code Form is subject to the terms of the Mozilla Public |
| 2 # License, v. 2.0. If a copy of the MPL was not distributed with this | 2 # License, v. 2.0. If a copy of the MPL was not distributed with this |
| 3 # file, You can obtain one at http://mozilla.org/MPL/2.0/. | 3 # file, You can obtain one at http://mozilla.org/MPL/2.0/. |
| 4 | 4 |
| 5 import errno | 5 import errno |
| 6 import glob | |
| 6 import io | 7 import io |
| 7 import json | 8 import json |
| 8 import os | 9 import os |
| 9 import re | 10 import re |
| 10 from StringIO import StringIO | 11 from StringIO import StringIO |
| 11 import struct | 12 import struct |
| 13 import subprocess | |
| 12 import sys | 14 import sys |
| 13 import collections | |
| 14 import glob | |
| 15 | 15 |
| 16 from packager import (readMetadata, getDefaultFileName, getBuildVersion, | 16 from packager import (readMetadata, getDefaultFileName, getBuildVersion, |
| 17 getTemplate, Files) | 17 getTemplate, Files) |
| 18 | 18 |
| 19 defaultLocale = 'en_US' | 19 defaultLocale = 'en_US' |
| 20 | 20 |
| 21 | 21 |
| 22 def getIgnoredFiles(params): | 22 def getIgnoredFiles(params): |
| 23 return {'store.description'} | 23 return {'store.description'} |
| 24 | 24 |
| (...skipping 108 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 133 # Normalize JSON structure | 133 # Normalize JSON structure |
| 134 licenseComment = re.compile(r'/\*.*?\*/', re.S) | 134 licenseComment = re.compile(r'/\*.*?\*/', re.S) |
| 135 data = json.loads(re.sub(licenseComment, '', manifest, 1)) | 135 data = json.loads(re.sub(licenseComment, '', manifest, 1)) |
| 136 if '_dummy' in data: | 136 if '_dummy' in data: |
| 137 del data['_dummy'] | 137 del data['_dummy'] |
| 138 manifest = json.dumps(data, sort_keys=True, indent=2) | 138 manifest = json.dumps(data, sort_keys=True, indent=2) |
| 139 | 139 |
| 140 return manifest.encode('utf-8') | 140 return manifest.encode('utf-8') |
| 141 | 141 |
| 142 | 142 |
| 143 def convertJS(params, files): | |
| 144 output_files = collections.OrderedDict() | |
| 145 args = {} | |
| 146 | |
| 147 for item in params['metadata'].items('convert_js'): | |
| 148 name, value = item | |
| 149 filename, arg = re.search(r'^(.*?)(?:\[(.*)\])?$', name).groups() | |
| 150 if arg is None: | |
| 151 output_files[filename] = (value.split(), item.source) | |
| 152 else: | |
| 153 args.setdefault(filename, {})[arg] = value | |
| 154 | |
| 155 template = getTemplate('modules.js.tmpl') | |
| 156 | |
| 157 for filename, (input_files, origin) in output_files.iteritems(): | |
| 158 if '/' in filename and not files.isIncluded(filename): | |
| 159 continue | |
| 160 | |
| 161 current_args = args.get(filename, {}) | |
| 162 current_args['autoload'] = [module for module in | |
| 163 current_args.get('autoload', '').split(',') | |
| 164 if module != ''] | |
| 165 | |
| 166 base_dir = os.path.dirname(origin) | |
| 167 modules = [] | |
| 168 | |
| 169 for input_filename in input_files: | |
| 170 module_name = os.path.splitext(os.path.basename(input_filename))[0] | |
| 171 prefix = os.path.basename(os.path.dirname(input_filename)) | |
| 172 if prefix != 'lib': | |
| 173 module_name = '{}_{}'.format(prefix, module_name) | |
| 174 with open(os.path.join(base_dir, input_filename), 'r') as file: | |
| 175 modules.append((module_name, file.read().decode('utf-8'))) | |
| 176 files.pop(input_filename, None) | |
| 177 | |
| 178 files[filename] = template.render( | |
| 179 args=current_args, | |
| 180 basename=params['metadata'].get('general', 'basename'), | |
| 181 modules=modules, | |
| 182 type=params['type'], | |
| 183 version=params['metadata'].get('general', 'version') | |
| 184 ).encode('utf-8') | |
| 185 | |
| 186 | |
| 187 def toJson(data): | 143 def toJson(data): |
| 188 return json.dumps( | 144 return json.dumps( |
| 189 data, ensure_ascii=False, sort_keys=True, | 145 data, ensure_ascii=False, sort_keys=True, |
| 190 indent=2, separators=(',', ': ') | 146 indent=2, separators=(',', ': ') |
| 191 ).encode('utf-8') + '\n' | 147 ).encode('utf-8') + '\n' |
| 192 | 148 |
| 193 | 149 |
| 150 def create_bundles(params, files): | |
| 151 base_extension_path = params['baseDir'] | |
| 152 info_templates = { | |
| 153 'chrome': 'chromeInfo.js.tmpl', | |
| 154 'edge': 'edgeInfo.js.tmpl', | |
| 155 'gecko-webext': 'geckoInfo.js.tmpl' | |
| 156 } | |
| 157 | |
| 158 # Historically we didn't use relative paths when requiring modules, so in | |
| 159 # order for webpack to know where to find them we need to pass in a list of | |
| 160 # resolve paths. Going forward we should always use relative paths, once we | |
| 161 # do that consistently this can be removed. See issues 5760, 5761 and 5762. | |
| 162 resolve_paths = [os.path.join(base_extension_path, dir, 'lib') | |
| 163 for dir in ['', 'adblockpluscore', 'adblockplusui']] | |
| 164 | |
| 165 info_template = getTemplate(info_templates[params['type']]) | |
| 166 info_module = info_template.render( | |
| 167 basename=params['metadata'].get('general', 'basename'), | |
| 168 version=params['metadata'].get('general', 'version') | |
| 169 ).encode('utf-8') | |
| 170 | |
| 171 configuration = { | |
| 172 'bundles': [], | |
| 173 'extension_path': base_extension_path, | |
| 174 'info_module': info_module, | |
| 175 'resolve_paths': resolve_paths, | |
| 176 } | |
| 177 | |
| 178 for item in params['metadata'].items('bundles'): | |
| 179 name, value = item | |
| 180 base_item_path = os.path.dirname(item.source) | |
| 181 | |
| 182 bundle_file = os.path.relpath(os.path.join(base_item_path, name), | |
| 183 base_extension_path) | |
| 184 entry_files = [os.path.join(base_item_path, module_path) | |
| 185 for module_path in value.split()] | |
| 186 configuration['bundles'].append({ | |
| 187 'bundle_name': bundle_file, | |
| 188 'entry_points': entry_files, | |
| 189 }) | |
| 190 | |
| 191 cmd = ['node', os.path.join(os.path.dirname(__file__), 'webpack_runner.js')] | |
| 192 process = subprocess.Popen(cmd, stdout=subprocess.PIPE, | |
| 193 stdin=subprocess.PIPE) | |
| 194 output = process.communicate(input=toJson(configuration))[0] | |
| 195 if process.returncode != 0: | |
| 196 raise subprocess.CalledProcessError(process.returncode, cmd=cmd) | |
| 197 output = json.loads(output) | |
| 198 | |
| 199 # Clear the mapping for any files included in a bundle, to avoid them being | |
| 200 # duplicated in the build. | |
| 201 for to_ignore in output['included']: | |
| 202 if to_ignore in files: | |
| 203 del files[to_ignore] | |
|
Sebastian Noack
2017/10/12 23:23:28
You can use files.pop(to_ignore, None), then you d
kzar
2017/10/13 07:25:21
Yea we can do it that way too, Done.
| |
| 204 | |
| 205 for bundle in output['files']: | |
| 206 files[bundle] = output['files'][bundle].encode('utf-8') | |
| 207 | |
| 208 | |
| 194 def import_locales(params, files): | 209 def import_locales(params, files): |
| 195 for item in params['metadata'].items('import_locales'): | 210 for item in params['metadata'].items('import_locales'): |
| 196 filename, keys = item | 211 filename, keys = item |
| 197 for sourceFile in glob.glob(os.path.join(os.path.dirname(item.source), | 212 for sourceFile in glob.glob(os.path.join(os.path.dirname(item.source), |
| 198 *filename.split('/'))): | 213 *filename.split('/'))): |
| 199 locale = sourceFile.split(os.path.sep)[-2] | 214 locale = sourceFile.split(os.path.sep)[-2] |
| 200 targetFile = os.path.join('_locales', locale, 'messages.json') | 215 targetFile = os.path.join('_locales', locale, 'messages.json') |
| 201 data = json.loads(files.get(targetFile, '{}').decode('utf-8')) | 216 data = json.loads(files.get(targetFile, '{}').decode('utf-8')) |
| 202 | 217 |
| 203 try: | 218 try: |
| (...skipping 121 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 325 'metadata': metadata, | 340 'metadata': metadata, |
| 326 } | 341 } |
| 327 | 342 |
| 328 mapped = metadata.items('mapping') if metadata.has_section('mapping') else [ ] | 343 mapped = metadata.items('mapping') if metadata.has_section('mapping') else [ ] |
| 329 files = Files(getPackageFiles(params), getIgnoredFiles(params), | 344 files = Files(getPackageFiles(params), getIgnoredFiles(params), |
| 330 process=lambda path, data: processFile(path, data, params)) | 345 process=lambda path, data: processFile(path, data, params)) |
| 331 | 346 |
| 332 files.readMappedFiles(mapped) | 347 files.readMappedFiles(mapped) |
| 333 files.read(baseDir, skip=[opt for opt, _ in mapped]) | 348 files.read(baseDir, skip=[opt for opt, _ in mapped]) |
| 334 | 349 |
| 335 if metadata.has_section('convert_js'): | 350 if metadata.has_section('bundles'): |
| 336 convertJS(params, files) | 351 create_bundles(params, files) |
| 337 | 352 |
| 338 if metadata.has_section('preprocess'): | 353 if metadata.has_section('preprocess'): |
| 339 files.preprocess( | 354 files.preprocess( |
| 340 [f for f, _ in metadata.items('preprocess')], | 355 [f for f, _ in metadata.items('preprocess')], |
| 341 {'needsExt': True} | 356 {'needsExt': True} |
| 342 ) | 357 ) |
| 343 | 358 |
| 344 if metadata.has_section('import_locales'): | 359 if metadata.has_section('import_locales'): |
| 345 import_locales(params, files) | 360 import_locales(params, files) |
| 346 | 361 |
| (...skipping 12 matching lines...) Expand all Loading... | |
| 359 params, 'testIndex.html.tmpl', ('general', 'testScripts') | 374 params, 'testIndex.html.tmpl', ('general', 'testScripts') |
| 360 ) | 375 ) |
| 361 | 376 |
| 362 zipdata = files.zipToString() | 377 zipdata = files.zipToString() |
| 363 signature = None | 378 signature = None |
| 364 pubkey = None | 379 pubkey = None |
| 365 if keyFile != None: | 380 if keyFile != None: |
| 366 signature = signBinary(zipdata, keyFile) | 381 signature = signBinary(zipdata, keyFile) |
| 367 pubkey = getPublicKey(keyFile) | 382 pubkey = getPublicKey(keyFile) |
| 368 writePackage(outFile, pubkey, signature, zipdata) | 383 writePackage(outFile, pubkey, signature, zipdata) |
| OLD | NEW |