mirror of
				https://github.com/mozilla/gecko-dev.git
				synced 2025-11-04 10:18:41 +02:00 
			
		
		
		
	
		
			
				
	
	
		
			310 lines
		
	
	
	
		
			9 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable file
		
	
	
	
	
			
		
		
	
	
			310 lines
		
	
	
	
		
			9 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable file
		
	
	
	
	
#!/usr/bin/env python
 | 
						|
# Copyright 2017 The Chromium Authors. All rights reserved.
 | 
						|
# Use of this source code is governed by a BSD-style license that can be
 | 
						|
# found in the LICENSE file.
 | 
						|
 | 
						|
"""Find header files missing in GN.
 | 
						|
 | 
						|
This script gets all the header files from ninja_deps, which is from the true
 | 
						|
dependency generated by the compiler, and report if they don't exist in GN.
 | 
						|
"""
 | 
						|
 | 
						|
from __future__ import print_function
 | 
						|
 | 
						|
import argparse
 | 
						|
import json
 | 
						|
import os
 | 
						|
import re
 | 
						|
import shutil
 | 
						|
import subprocess
 | 
						|
import sys
 | 
						|
import tempfile
 | 
						|
from multiprocessing import Process, Queue
 | 
						|
 | 
						|
SRC_DIR = os.path.abspath(
 | 
						|
    os.path.join(os.path.abspath(os.path.dirname(__file__)), os.path.pardir))
 | 
						|
DEPOT_TOOLS_DIR = os.path.join(SRC_DIR, 'third_party', 'depot_tools')
 | 
						|
 | 
						|
 | 
						|
def GetHeadersFromNinja(out_dir, skip_obj, q):
 | 
						|
  """Return all the header files from ninja_deps"""
 | 
						|
 | 
						|
  def NinjaSource():
 | 
						|
    cmd = [os.path.join(DEPOT_TOOLS_DIR, 'ninja'), '-C', out_dir, '-t', 'deps']
 | 
						|
    # A negative bufsize means to use the system default, which usually
 | 
						|
    # means fully buffered.
 | 
						|
    popen = subprocess.Popen(cmd, stdout=subprocess.PIPE, bufsize=-1)
 | 
						|
    for line in iter(popen.stdout.readline, ''):
 | 
						|
      yield line.rstrip()
 | 
						|
 | 
						|
    popen.stdout.close()
 | 
						|
    return_code = popen.wait()
 | 
						|
    if return_code:
 | 
						|
      raise subprocess.CalledProcessError(return_code, cmd)
 | 
						|
 | 
						|
  ans, err = set(), None
 | 
						|
  try:
 | 
						|
    ans = ParseNinjaDepsOutput(NinjaSource(), out_dir, skip_obj)
 | 
						|
  except Exception as e:
 | 
						|
    err = str(e)
 | 
						|
  q.put((ans, err))
 | 
						|
 | 
						|
 | 
						|
def ParseNinjaDepsOutput(ninja_out, out_dir, skip_obj):
 | 
						|
  """Parse ninja output and get the header files"""
 | 
						|
  all_headers = {}
 | 
						|
 | 
						|
  # Ninja always uses "/", even on Windows.
 | 
						|
  prefix = '../../'
 | 
						|
 | 
						|
  is_valid = False
 | 
						|
  obj_file = ''
 | 
						|
  for line in ninja_out:
 | 
						|
    if line.startswith('    '):
 | 
						|
      if not is_valid:
 | 
						|
        continue
 | 
						|
      if line.endswith('.h') or line.endswith('.hh'):
 | 
						|
        f = line.strip()
 | 
						|
        if f.startswith(prefix):
 | 
						|
          f = f[6:]  # Remove the '../../' prefix
 | 
						|
          # build/ only contains build-specific files like build_config.h
 | 
						|
          # and buildflag.h, and system header files, so they should be
 | 
						|
          # skipped.
 | 
						|
          if f.startswith(out_dir) or f.startswith('out'):
 | 
						|
            continue
 | 
						|
          if not f.startswith('build'):
 | 
						|
            all_headers.setdefault(f, [])
 | 
						|
            if not skip_obj:
 | 
						|
              all_headers[f].append(obj_file)
 | 
						|
    else:
 | 
						|
      is_valid = line.endswith('(VALID)')
 | 
						|
      obj_file = line.split(':')[0]
 | 
						|
 | 
						|
  return all_headers
 | 
						|
 | 
						|
 | 
						|
def GetHeadersFromGN(out_dir, q):
 | 
						|
  """Return all the header files from GN"""
 | 
						|
 | 
						|
  tmp = None
 | 
						|
  ans, err = set(), None
 | 
						|
  try:
 | 
						|
    # Argument |dir| is needed to make sure it's on the same drive on Windows.
 | 
						|
    # dir='' means dir='.', but doesn't introduce an unneeded prefix.
 | 
						|
    tmp = tempfile.mkdtemp(dir='')
 | 
						|
    shutil.copy2(os.path.join(out_dir, 'args.gn'),
 | 
						|
                 os.path.join(tmp, 'args.gn'))
 | 
						|
    # Do "gn gen" in a temp dir to prevent dirtying |out_dir|.
 | 
						|
    gn_exe = 'gn.bat' if sys.platform == 'win32' else 'gn'
 | 
						|
    subprocess.check_call([
 | 
						|
        os.path.join(DEPOT_TOOLS_DIR, gn_exe), 'gen', tmp, '--ide=json', '-q'])
 | 
						|
    gn_json = json.load(open(os.path.join(tmp, 'project.json')))
 | 
						|
    ans = ParseGNProjectJSON(gn_json, out_dir, tmp)
 | 
						|
  except Exception as e:
 | 
						|
    err = str(e)
 | 
						|
  finally:
 | 
						|
    if tmp:
 | 
						|
      shutil.rmtree(tmp)
 | 
						|
  q.put((ans, err))
 | 
						|
 | 
						|
 | 
						|
def ParseGNProjectJSON(gn, out_dir, tmp_out):
 | 
						|
  """Parse GN output and get the header files"""
 | 
						|
  all_headers = set()
 | 
						|
 | 
						|
  for _target, properties in gn['targets'].iteritems():
 | 
						|
    sources = properties.get('sources', [])
 | 
						|
    public = properties.get('public', [])
 | 
						|
    # Exclude '"public": "*"'.
 | 
						|
    if type(public) is list:
 | 
						|
      sources += public
 | 
						|
    for f in sources:
 | 
						|
      if f.endswith('.h') or f.endswith('.hh'):
 | 
						|
        if f.startswith('//'):
 | 
						|
          f = f[2:]  # Strip the '//' prefix.
 | 
						|
          if f.startswith(tmp_out):
 | 
						|
            f = out_dir + f[len(tmp_out):]
 | 
						|
          all_headers.add(f)
 | 
						|
 | 
						|
  return all_headers
 | 
						|
 | 
						|
 | 
						|
def GetDepsPrefixes(q):
 | 
						|
  """Return all the folders controlled by DEPS file"""
 | 
						|
  prefixes, err = set(), None
 | 
						|
  try:
 | 
						|
    gclient_exe = 'gclient.bat' if sys.platform == 'win32' else 'gclient'
 | 
						|
    gclient_out = subprocess.check_output([
 | 
						|
        os.path.join(DEPOT_TOOLS_DIR, gclient_exe),
 | 
						|
        'recurse', '--no-progress', '-j1',
 | 
						|
        'python', '-c', 'import os;print os.environ["GCLIENT_DEP_PATH"]'],
 | 
						|
        universal_newlines=True)
 | 
						|
    for i in gclient_out.split('\n'):
 | 
						|
      if i.startswith('src/'):
 | 
						|
        i = i[4:]
 | 
						|
        prefixes.add(i)
 | 
						|
  except Exception as e:
 | 
						|
    err = str(e)
 | 
						|
  q.put((prefixes, err))
 | 
						|
 | 
						|
 | 
						|
def IsBuildClean(out_dir):
 | 
						|
  cmd = [os.path.join(DEPOT_TOOLS_DIR, 'ninja'), '-C', out_dir, '-n']
 | 
						|
  try:
 | 
						|
    out = subprocess.check_output(cmd)
 | 
						|
    return 'no work to do.' in out
 | 
						|
  except Exception as e:
 | 
						|
    print(e)
 | 
						|
    return False
 | 
						|
 | 
						|
def ParseWhiteList(whitelist):
 | 
						|
  out = set()
 | 
						|
  for line in whitelist.split('\n'):
 | 
						|
    line = re.sub(r'#.*', '', line).strip()
 | 
						|
    if line:
 | 
						|
      out.add(line)
 | 
						|
  return out
 | 
						|
 | 
						|
 | 
						|
def FilterOutDepsedRepo(files, deps):
 | 
						|
  return {f for f in files if not any(f.startswith(d) for d in deps)}
 | 
						|
 | 
						|
 | 
						|
def GetNonExistingFiles(lst):
 | 
						|
  out = set()
 | 
						|
  for f in lst:
 | 
						|
    if not os.path.isfile(f):
 | 
						|
      out.add(f)
 | 
						|
  return out
 | 
						|
 | 
						|
 | 
						|
def main():
 | 
						|
 | 
						|
  def DumpJson(data):
 | 
						|
    if args.json:
 | 
						|
      with open(args.json, 'w') as f:
 | 
						|
        json.dump(data, f)
 | 
						|
 | 
						|
  def PrintError(msg):
 | 
						|
    DumpJson([])
 | 
						|
    parser.error(msg)
 | 
						|
 | 
						|
  parser = argparse.ArgumentParser(description='''
 | 
						|
      NOTE: Use ninja to build all targets in OUT_DIR before running
 | 
						|
      this script.''')
 | 
						|
  parser.add_argument('--out-dir', metavar='OUT_DIR', default='out/Release',
 | 
						|
                      help='output directory of the build')
 | 
						|
  parser.add_argument('--json',
 | 
						|
                      help='JSON output filename for missing headers')
 | 
						|
  parser.add_argument('--whitelist', help='file containing whitelist')
 | 
						|
  parser.add_argument('--skip-dirty-check', action='store_true',
 | 
						|
                      help='skip checking whether the build is dirty')
 | 
						|
  parser.add_argument('--verbose', action='store_true',
 | 
						|
                      help='print more diagnostic info')
 | 
						|
 | 
						|
  args, _extras = parser.parse_known_args()
 | 
						|
 | 
						|
  if not os.path.isdir(args.out_dir):
 | 
						|
    parser.error('OUT_DIR "%s" does not exist.' % args.out_dir)
 | 
						|
 | 
						|
  if not args.skip_dirty_check and not IsBuildClean(args.out_dir):
 | 
						|
    dirty_msg = 'OUT_DIR looks dirty. You need to build all there.'
 | 
						|
    if args.json:
 | 
						|
      # Assume running on the bots. Silently skip this step.
 | 
						|
      # This is possible because "analyze" step can be wrong due to
 | 
						|
      # underspecified header files. See crbug.com/725877
 | 
						|
      print(dirty_msg)
 | 
						|
      DumpJson([])
 | 
						|
      return 0
 | 
						|
    else:
 | 
						|
      # Assume running interactively.
 | 
						|
      parser.error(dirty_msg)
 | 
						|
 | 
						|
  d_q = Queue()
 | 
						|
  d_p = Process(target=GetHeadersFromNinja, args=(args.out_dir, True, d_q,))
 | 
						|
  d_p.start()
 | 
						|
 | 
						|
  gn_q = Queue()
 | 
						|
  gn_p = Process(target=GetHeadersFromGN, args=(args.out_dir, gn_q,))
 | 
						|
  gn_p.start()
 | 
						|
 | 
						|
  deps_q = Queue()
 | 
						|
  deps_p = Process(target=GetDepsPrefixes, args=(deps_q,))
 | 
						|
  deps_p.start()
 | 
						|
 | 
						|
  d, d_err = d_q.get()
 | 
						|
  gn, gn_err = gn_q.get()
 | 
						|
  missing = set(d.keys()) - gn
 | 
						|
  nonexisting = GetNonExistingFiles(gn)
 | 
						|
 | 
						|
  deps, deps_err = deps_q.get()
 | 
						|
  missing = FilterOutDepsedRepo(missing, deps)
 | 
						|
  nonexisting = FilterOutDepsedRepo(nonexisting, deps)
 | 
						|
 | 
						|
  d_p.join()
 | 
						|
  gn_p.join()
 | 
						|
  deps_p.join()
 | 
						|
 | 
						|
  if d_err:
 | 
						|
    PrintError(d_err)
 | 
						|
  if gn_err:
 | 
						|
    PrintError(gn_err)
 | 
						|
  if deps_err:
 | 
						|
    PrintError(deps_err)
 | 
						|
  if len(GetNonExistingFiles(d)) > 0:
 | 
						|
    print('Non-existing files in ninja deps:', GetNonExistingFiles(d))
 | 
						|
    PrintError('Found non-existing files in ninja deps. You should ' +
 | 
						|
               'build all in OUT_DIR.')
 | 
						|
  if len(d) == 0:
 | 
						|
    PrintError('OUT_DIR looks empty. You should build all there.')
 | 
						|
  if any((('/gen/' in i) for i in nonexisting)):
 | 
						|
    PrintError('OUT_DIR looks wrong. You should build all there.')
 | 
						|
 | 
						|
  if args.whitelist:
 | 
						|
    whitelist = ParseWhiteList(open(args.whitelist).read())
 | 
						|
    missing -= whitelist
 | 
						|
    nonexisting -= whitelist
 | 
						|
 | 
						|
  missing = sorted(missing)
 | 
						|
  nonexisting = sorted(nonexisting)
 | 
						|
 | 
						|
  DumpJson(sorted(missing + nonexisting))
 | 
						|
 | 
						|
  if len(missing) == 0 and len(nonexisting) == 0:
 | 
						|
    return 0
 | 
						|
 | 
						|
  if len(missing) > 0:
 | 
						|
    print('\nThe following files should be included in gn files:')
 | 
						|
    for i in missing:
 | 
						|
      print(i)
 | 
						|
 | 
						|
  if len(nonexisting) > 0:
 | 
						|
    print('\nThe following non-existing files should be removed from gn files:')
 | 
						|
    for i in nonexisting:
 | 
						|
      print(i)
 | 
						|
 | 
						|
  if args.verbose:
 | 
						|
    # Only get detailed obj dependency here since it is slower.
 | 
						|
    GetHeadersFromNinja(args.out_dir, False, d_q)
 | 
						|
    d, d_err = d_q.get()
 | 
						|
    print('\nDetailed dependency info:')
 | 
						|
    for f in missing:
 | 
						|
      print(f)
 | 
						|
      for cc in d[f]:
 | 
						|
        print('  ', cc)
 | 
						|
 | 
						|
    print('\nMissing headers sorted by number of affected object files:')
 | 
						|
    count = {k: len(v) for (k, v) in d.iteritems()}
 | 
						|
    for f in sorted(count, key=count.get, reverse=True):
 | 
						|
      if f in missing:
 | 
						|
        print(count[f], f)
 | 
						|
 | 
						|
  if args.json:
 | 
						|
    # Assume running on the bots. Temporarily return 0 before
 | 
						|
    # https://crbug.com/937847 is fixed.
 | 
						|
    return 0
 | 
						|
  return 1
 | 
						|
 | 
						|
 | 
						|
if __name__ == '__main__':
 | 
						|
  sys.exit(main())
 |