forked from mirrors/gecko-dev
		
	 3f4ac974d3
			
		
	
	
		3f4ac974d3
		
	
	
	
	
		
			
			yaml.load() can evaluate arbitrary Python code via syntax such as `!!python/object/apply:os.system`. Seriously. Let's switch taskgraph to yaml.safe_load(), which is reasonable about limiting magic. Differential Revision: https://phabricator.services.mozilla.com/D1736
		
			
				
	
	
		
			301 lines
		
	
	
	
		
			10 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			301 lines
		
	
	
	
		
			10 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # This Source Code Form is subject to the terms of the Mozilla Public
 | |
| # License, v. 2.0. If a copy of the MPL was not distributed with this
 | |
| # file, You can obtain one at http://mozilla.org/MPL/2.0/.
 | |
| 
 | |
| from __future__ import absolute_import, print_function, unicode_literals
 | |
| 
 | |
| import hashlib
 | |
| import json
 | |
| import os
 | |
| import re
 | |
| import requests_unixsocket
 | |
| import sys
 | |
| import urllib
 | |
| import urlparse
 | |
| import yaml
 | |
| 
 | |
| from mozbuild.util import memoize
 | |
| from mozpack.files import GeneratedFile
 | |
| from mozpack.archive import (
 | |
|     create_tar_gz_from_files,
 | |
| )
 | |
| from .. import GECKO
 | |
| 
 | |
| 
 | |
| IMAGE_DIR = os.path.join(GECKO, 'taskcluster', 'docker')
 | |
| 
 | |
| 
 | |
| def docker_url(path, **kwargs):
 | |
|     docker_socket = os.environ.get('DOCKER_SOCKET', '/var/run/docker.sock')
 | |
|     return urlparse.urlunparse((
 | |
|         'http+unix',
 | |
|         urllib.quote(docker_socket, safe=''),
 | |
|         path,
 | |
|         '',
 | |
|         urllib.urlencode(kwargs),
 | |
|         ''))
 | |
| 
 | |
| 
 | |
| def post_to_docker(tar, api_path, **kwargs):
 | |
|     """POSTs a tar file to a given docker API path.
 | |
| 
 | |
|     The tar argument can be anything that can be passed to requests.post()
 | |
|     as data (e.g. iterator or file object).
 | |
|     The extra keyword arguments are passed as arguments to the docker API.
 | |
|     """
 | |
|     req = requests_unixsocket.Session().post(
 | |
|         docker_url(api_path, **kwargs),
 | |
|         data=tar,
 | |
|         stream=True,
 | |
|         headers={'Content-Type': 'application/x-tar'},
 | |
|     )
 | |
|     if req.status_code != 200:
 | |
|         message = req.json().get('message')
 | |
|         if not message:
 | |
|             message = 'docker API returned HTTP code {}'.format(
 | |
|                 req.status_code)
 | |
|         raise Exception(message)
 | |
|     status_line = {}
 | |
| 
 | |
|     buf = b''
 | |
|     for content in req.iter_content(chunk_size=None):
 | |
|         if not content:
 | |
|             continue
 | |
|         # Sometimes, a chunk of content is not a complete json, so we cumulate
 | |
|         # with leftovers from previous iterations.
 | |
|         buf += content
 | |
|         try:
 | |
|             data = json.loads(buf)
 | |
|         except Exception:
 | |
|             continue
 | |
|         buf = b''
 | |
|         # data is sometimes an empty dict.
 | |
|         if not data:
 | |
|             continue
 | |
|         # Mimick how docker itself presents the output. This code was tested
 | |
|         # with API version 1.18 and 1.26.
 | |
|         if 'status' in data:
 | |
|             if 'id' in data:
 | |
|                 if sys.stderr.isatty():
 | |
|                     total_lines = len(status_line)
 | |
|                     line = status_line.setdefault(data['id'], total_lines)
 | |
|                     n = total_lines - line
 | |
|                     if n > 0:
 | |
|                         # Move the cursor up n lines.
 | |
|                         sys.stderr.write('\033[{}A'.format(n))
 | |
|                     # Clear line and move the cursor to the beginning of it.
 | |
|                     sys.stderr.write('\033[2K\r')
 | |
|                     sys.stderr.write('{}: {} {}\n'.format(
 | |
|                         data['id'], data['status'], data.get('progress', '')))
 | |
|                     if n > 1:
 | |
|                         # Move the cursor down n - 1 lines, which, considering
 | |
|                         # the carriage return on the last write, gets us back
 | |
|                         # where we started.
 | |
|                         sys.stderr.write('\033[{}B'.format(n - 1))
 | |
|                 else:
 | |
|                     status = status_line.get(data['id'])
 | |
|                     # Only print status changes.
 | |
|                     if status != data['status']:
 | |
|                         sys.stderr.write('{}: {}\n'.format(data['id'], data['status']))
 | |
|                         status_line[data['id']] = data['status']
 | |
|             else:
 | |
|                 status_line = {}
 | |
|                 sys.stderr.write('{}\n'.format(data['status']))
 | |
|         elif 'stream' in data:
 | |
|             sys.stderr.write(data['stream'])
 | |
|         elif 'error' in data:
 | |
|             raise Exception(data['error'])
 | |
|         else:
 | |
|             raise NotImplementedError(repr(data))
 | |
|         sys.stderr.flush()
 | |
| 
 | |
| 
 | |
| def docker_image(name, by_tag=False):
 | |
|     '''
 | |
|         Resolve in-tree prebuilt docker image to ``<registry>/<repository>@sha256:<digest>``,
 | |
|         or ``<registry>/<repository>:<tag>`` if `by_tag` is `True`.
 | |
|     '''
 | |
|     try:
 | |
|         with open(os.path.join(IMAGE_DIR, name, 'REGISTRY')) as f:
 | |
|             registry = f.read().strip()
 | |
|     except IOError:
 | |
|         with open(os.path.join(IMAGE_DIR, 'REGISTRY')) as f:
 | |
|             registry = f.read().strip()
 | |
| 
 | |
|     if not by_tag:
 | |
|         hashfile = os.path.join(IMAGE_DIR, name, 'HASH')
 | |
|         try:
 | |
|             with open(hashfile) as f:
 | |
|                 return '{}/{}@{}'.format(registry, name, f.read().strip())
 | |
|         except IOError:
 | |
|             raise Exception('Failed to read HASH file {}'.format(hashfile))
 | |
| 
 | |
|     try:
 | |
|         with open(os.path.join(IMAGE_DIR, name, 'VERSION')) as f:
 | |
|             tag = f.read().strip()
 | |
|     except IOError:
 | |
|         tag = 'latest'
 | |
|     return '{}/{}:{}'.format(registry, name, tag)
 | |
| 
 | |
| 
 | |
| class VoidWriter(object):
 | |
|     """A file object with write capabilities that does nothing with the written
 | |
|     data."""
 | |
|     def write(self, buf):
 | |
|         pass
 | |
| 
 | |
| 
 | |
| def generate_context_hash(topsrcdir, image_path, image_name, args=None):
 | |
|     """Generates a sha256 hash for context directory used to build an image."""
 | |
| 
 | |
|     return stream_context_tar(topsrcdir, image_path, VoidWriter(), image_name, args)
 | |
| 
 | |
| 
 | |
| class HashingWriter(object):
 | |
|     """A file object with write capabilities that hashes the written data at
 | |
|     the same time it passes down to a real file object."""
 | |
|     def __init__(self, writer):
 | |
|         self._hash = hashlib.sha256()
 | |
|         self._writer = writer
 | |
| 
 | |
|     def write(self, buf):
 | |
|         self._hash.update(buf)
 | |
|         self._writer.write(buf)
 | |
| 
 | |
|     def hexdigest(self):
 | |
|         return self._hash.hexdigest()
 | |
| 
 | |
| 
 | |
| def create_context_tar(topsrcdir, context_dir, out_path, prefix, args=None):
 | |
|     """Create a context tarball.
 | |
| 
 | |
|     A directory ``context_dir`` containing a Dockerfile will be assembled into
 | |
|     a gzipped tar file at ``out_path``. Files inside the archive will be
 | |
|     prefixed by directory ``prefix``.
 | |
| 
 | |
|     We also scan the source Dockerfile for special syntax that influences
 | |
|     context generation.
 | |
| 
 | |
|     If a line in the Dockerfile has the form ``# %include <path>``,
 | |
|     the relative path specified on that line will be matched against
 | |
|     files in the source repository and added to the context under the
 | |
|     path ``topsrcdir/``. If an entry is a directory, we add all files
 | |
|     under that directory.
 | |
| 
 | |
|     If a line in the Dockerfile has the form ``# %ARG <name>``, occurrences of
 | |
|     the string ``$<name>`` in subsequent lines are replaced with the value
 | |
|     found in the ``args`` argument. Exception: this doesn't apply to VOLUME
 | |
|     definitions.
 | |
| 
 | |
|     Returns the SHA-256 hex digest of the created archive.
 | |
|     """
 | |
|     with open(out_path, 'wb') as fh:
 | |
|         return stream_context_tar(topsrcdir, context_dir, fh, prefix, args)
 | |
| 
 | |
| 
 | |
| def stream_context_tar(topsrcdir, context_dir, out_file, prefix, args=None):
 | |
|     """Like create_context_tar, but streams the tar file to the `out_file` file
 | |
|     object."""
 | |
|     archive_files = {}
 | |
|     replace = []
 | |
| 
 | |
|     for root, dirs, files in os.walk(context_dir):
 | |
|         for f in files:
 | |
|             source_path = os.path.join(root, f)
 | |
|             rel = source_path[len(context_dir) + 1:]
 | |
|             archive_path = os.path.join(prefix, rel)
 | |
|             archive_files[archive_path] = source_path
 | |
| 
 | |
|     # Parse Dockerfile for special syntax of extra files to include.
 | |
|     content = []
 | |
|     with open(os.path.join(context_dir, 'Dockerfile'), 'rb') as fh:
 | |
|         for line in fh:
 | |
|             if line.startswith('# %ARG'):
 | |
|                 p = line[len('# %ARG '):].strip()
 | |
|                 if not args or p not in args:
 | |
|                     raise Exception('missing argument: {}'.format(p))
 | |
|                 replace.append((re.compile(r'\${}\b'.format(p)),
 | |
|                                 args[p].encode('ascii')))
 | |
|                 continue
 | |
| 
 | |
|             for regexp, s in replace:
 | |
|                 line = re.sub(regexp, s, line)
 | |
| 
 | |
|             content.append(line)
 | |
| 
 | |
|             if not line.startswith('# %include'):
 | |
|                 continue
 | |
| 
 | |
|             p = line[len('# %include '):].strip()
 | |
|             if os.path.isabs(p):
 | |
|                 raise Exception('extra include path cannot be absolute: %s' % p)
 | |
| 
 | |
|             fs_path = os.path.normpath(os.path.join(topsrcdir, p))
 | |
|             # Check for filesystem traversal exploits.
 | |
|             if not fs_path.startswith(topsrcdir):
 | |
|                 raise Exception('extra include path outside topsrcdir: %s' % p)
 | |
| 
 | |
|             if not os.path.exists(fs_path):
 | |
|                 raise Exception('extra include path does not exist: %s' % p)
 | |
| 
 | |
|             if os.path.isdir(fs_path):
 | |
|                 for root, dirs, files in os.walk(fs_path):
 | |
|                     for f in files:
 | |
|                         source_path = os.path.join(root, f)
 | |
|                         rel = source_path[len(fs_path) + 1:]
 | |
|                         archive_path = os.path.join(prefix, 'topsrcdir', p, rel)
 | |
|                         archive_files[archive_path] = source_path
 | |
|             else:
 | |
|                 archive_path = os.path.join(prefix, 'topsrcdir', p)
 | |
|                 archive_files[archive_path] = fs_path
 | |
| 
 | |
|     archive_files[os.path.join(prefix, 'Dockerfile')] = \
 | |
|         GeneratedFile(b''.join(content))
 | |
| 
 | |
|     writer = HashingWriter(out_file)
 | |
|     create_tar_gz_from_files(writer, archive_files, '%s.tar.gz' % prefix)
 | |
|     return writer.hexdigest()
 | |
| 
 | |
| 
 | |
| @memoize
 | |
| def image_paths():
 | |
|     """Return a map of image name to paths containing their Dockerfile.
 | |
|     """
 | |
|     with open(os.path.join(GECKO, 'taskcluster', 'ci', 'docker-image',
 | |
|                            'kind.yml')) as fh:
 | |
|         config = yaml.safe_load(fh)
 | |
|         return {
 | |
|             k: os.path.join(IMAGE_DIR, v.get('definition', k))
 | |
|             for k, v in config['jobs'].items()
 | |
|         }
 | |
| 
 | |
| 
 | |
| def image_path(name):
 | |
|     paths = image_paths()
 | |
|     if name in paths:
 | |
|         return paths[name]
 | |
|     return os.path.join(IMAGE_DIR, name)
 | |
| 
 | |
| 
 | |
| @memoize
 | |
| def parse_volumes(image):
 | |
|     """Parse VOLUME entries from a Dockerfile for an image."""
 | |
|     volumes = set()
 | |
| 
 | |
|     path = image_path(image)
 | |
| 
 | |
|     with open(os.path.join(path, 'Dockerfile'), 'rb') as fh:
 | |
|         for line in fh:
 | |
|             line = line.strip()
 | |
|             # We assume VOLUME definitions don't use %ARGS.
 | |
|             if not line.startswith(b'VOLUME '):
 | |
|                 continue
 | |
| 
 | |
|             v = line.split(None, 1)[1]
 | |
|             if v.startswith(b'['):
 | |
|                 raise ValueError('cannot parse array syntax for VOLUME; '
 | |
|                                  'convert to multiple entries')
 | |
| 
 | |
|             volumes |= set(v.split())
 | |
| 
 | |
|     return volumes
 |