mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-24 14:26:09 +03:00
305 lines
8.8 KiB
Python
305 lines
8.8 KiB
Python
|
#!/usr/bin/env python
|
||
|
# Copyright 2017 The Chromium Authors. All rights reserved.
|
||
|
# Use of this source code is governed by a BSD-style license that can be
|
||
|
# found in the LICENSE file.
|
||
|
|
||
|
"""Find header files missing in GN.
|
||
|
|
||
|
This script gets all the header files from ninja_deps, which is from the true
|
||
|
dependency generated by the compiler, and report if they don't exist in GN.
|
||
|
"""
|
||
|
|
||
|
import argparse
|
||
|
import json
|
||
|
import os
|
||
|
import re
|
||
|
import shutil
|
||
|
import subprocess
|
||
|
import sys
|
||
|
import tempfile
|
||
|
from multiprocessing import Process, Queue
|
||
|
|
||
|
SRC_DIR = os.path.abspath(
|
||
|
os.path.join(os.path.abspath(os.path.dirname(__file__)), os.path.pardir))
|
||
|
DEPOT_TOOLS_DIR = os.path.join(SRC_DIR, 'third_party', 'depot_tools')
|
||
|
|
||
|
|
||
|
def GetHeadersFromNinja(out_dir, skip_obj, q):
|
||
|
"""Return all the header files from ninja_deps"""
|
||
|
|
||
|
def NinjaSource():
|
||
|
cmd = [os.path.join(DEPOT_TOOLS_DIR, 'ninja'), '-C', out_dir, '-t', 'deps']
|
||
|
# A negative bufsize means to use the system default, which usually
|
||
|
# means fully buffered.
|
||
|
popen = subprocess.Popen(cmd, stdout=subprocess.PIPE, bufsize=-1)
|
||
|
for line in iter(popen.stdout.readline, ''):
|
||
|
yield line.rstrip()
|
||
|
|
||
|
popen.stdout.close()
|
||
|
return_code = popen.wait()
|
||
|
if return_code:
|
||
|
raise subprocess.CalledProcessError(return_code, cmd)
|
||
|
|
||
|
ans, err = set(), None
|
||
|
try:
|
||
|
ans = ParseNinjaDepsOutput(NinjaSource(), out_dir, skip_obj)
|
||
|
except Exception as e:
|
||
|
err = str(e)
|
||
|
q.put((ans, err))
|
||
|
|
||
|
|
||
|
def ParseNinjaDepsOutput(ninja_out, out_dir, skip_obj):
|
||
|
"""Parse ninja output and get the header files"""
|
||
|
all_headers = {}
|
||
|
|
||
|
# Ninja always uses "/", even on Windows.
|
||
|
prefix = '../../'
|
||
|
|
||
|
is_valid = False
|
||
|
obj_file = ''
|
||
|
for line in ninja_out:
|
||
|
if line.startswith(' '):
|
||
|
if not is_valid:
|
||
|
continue
|
||
|
if line.endswith('.h') or line.endswith('.hh'):
|
||
|
f = line.strip()
|
||
|
if f.startswith(prefix):
|
||
|
f = f[6:] # Remove the '../../' prefix
|
||
|
# build/ only contains build-specific files like build_config.h
|
||
|
# and buildflag.h, and system header files, so they should be
|
||
|
# skipped.
|
||
|
if f.startswith(out_dir) or f.startswith('out'):
|
||
|
continue
|
||
|
if not f.startswith('build'):
|
||
|
all_headers.setdefault(f, [])
|
||
|
if not skip_obj:
|
||
|
all_headers[f].append(obj_file)
|
||
|
else:
|
||
|
is_valid = line.endswith('(VALID)')
|
||
|
obj_file = line.split(':')[0]
|
||
|
|
||
|
return all_headers
|
||
|
|
||
|
|
||
|
def GetHeadersFromGN(out_dir, q):
|
||
|
"""Return all the header files from GN"""
|
||
|
|
||
|
tmp = None
|
||
|
ans, err = set(), None
|
||
|
try:
|
||
|
# Argument |dir| is needed to make sure it's on the same drive on Windows.
|
||
|
# dir='' means dir='.', but doesn't introduce an unneeded prefix.
|
||
|
tmp = tempfile.mkdtemp(dir='')
|
||
|
shutil.copy2(os.path.join(out_dir, 'args.gn'),
|
||
|
os.path.join(tmp, 'args.gn'))
|
||
|
# Do "gn gen" in a temp dir to prevent dirtying |out_dir|.
|
||
|
gn_exe = 'gn.bat' if sys.platform == 'win32' else 'gn'
|
||
|
subprocess.check_call([
|
||
|
os.path.join(DEPOT_TOOLS_DIR, gn_exe), 'gen', tmp, '--ide=json', '-q'])
|
||
|
gn_json = json.load(open(os.path.join(tmp, 'project.json')))
|
||
|
ans = ParseGNProjectJSON(gn_json, out_dir, tmp)
|
||
|
except Exception as e:
|
||
|
err = str(e)
|
||
|
finally:
|
||
|
if tmp:
|
||
|
shutil.rmtree(tmp)
|
||
|
q.put((ans, err))
|
||
|
|
||
|
|
||
|
def ParseGNProjectJSON(gn, out_dir, tmp_out):
|
||
|
"""Parse GN output and get the header files"""
|
||
|
all_headers = set()
|
||
|
|
||
|
for _target, properties in gn['targets'].iteritems():
|
||
|
sources = properties.get('sources', [])
|
||
|
public = properties.get('public', [])
|
||
|
# Exclude '"public": "*"'.
|
||
|
if type(public) is list:
|
||
|
sources += public
|
||
|
for f in sources:
|
||
|
if f.endswith('.h') or f.endswith('.hh'):
|
||
|
if f.startswith('//'):
|
||
|
f = f[2:] # Strip the '//' prefix.
|
||
|
if f.startswith(tmp_out):
|
||
|
f = out_dir + f[len(tmp_out):]
|
||
|
all_headers.add(f)
|
||
|
|
||
|
return all_headers
|
||
|
|
||
|
|
||
|
def GetDepsPrefixes(q):
|
||
|
"""Return all the folders controlled by DEPS file"""
|
||
|
prefixes, err = set(), None
|
||
|
try:
|
||
|
gclient_exe = 'gclient.bat' if sys.platform == 'win32' else 'gclient'
|
||
|
gclient_out = subprocess.check_output([
|
||
|
os.path.join(DEPOT_TOOLS_DIR, gclient_exe),
|
||
|
'recurse', '--no-progress', '-j1',
|
||
|
'python', '-c', 'import os;print os.environ["GCLIENT_DEP_PATH"]'],
|
||
|
universal_newlines=True)
|
||
|
for i in gclient_out.split('\n'):
|
||
|
if i.startswith('src/'):
|
||
|
i = i[4:]
|
||
|
prefixes.add(i)
|
||
|
except Exception as e:
|
||
|
err = str(e)
|
||
|
q.put((prefixes, err))
|
||
|
|
||
|
|
||
|
def IsBuildClean(out_dir):
|
||
|
cmd = [os.path.join(DEPOT_TOOLS_DIR, 'ninja'), '-C', out_dir, '-n']
|
||
|
try:
|
||
|
out = subprocess.check_output(cmd)
|
||
|
return 'no work to do.' in out
|
||
|
except Exception as e:
|
||
|
print e
|
||
|
return False
|
||
|
|
||
|
def ParseWhiteList(whitelist):
|
||
|
out = set()
|
||
|
for line in whitelist.split('\n'):
|
||
|
line = re.sub(r'#.*', '', line).strip()
|
||
|
if line:
|
||
|
out.add(line)
|
||
|
return out
|
||
|
|
||
|
|
||
|
def FilterOutDepsedRepo(files, deps):
|
||
|
return {f for f in files if not any(f.startswith(d) for d in deps)}
|
||
|
|
||
|
|
||
|
def GetNonExistingFiles(lst):
|
||
|
out = set()
|
||
|
for f in lst:
|
||
|
if not os.path.isfile(f):
|
||
|
out.add(f)
|
||
|
return out
|
||
|
|
||
|
|
||
|
def main():
|
||
|
|
||
|
def DumpJson(data):
|
||
|
if args.json:
|
||
|
with open(args.json, 'w') as f:
|
||
|
json.dump(data, f)
|
||
|
|
||
|
def PrintError(msg):
|
||
|
DumpJson([])
|
||
|
parser.error(msg)
|
||
|
|
||
|
parser = argparse.ArgumentParser(description='''
|
||
|
NOTE: Use ninja to build all targets in OUT_DIR before running
|
||
|
this script.''')
|
||
|
parser.add_argument('--out-dir', metavar='OUT_DIR', default='out/Release',
|
||
|
help='output directory of the build')
|
||
|
parser.add_argument('--json',
|
||
|
help='JSON output filename for missing headers')
|
||
|
parser.add_argument('--whitelist', help='file containing whitelist')
|
||
|
parser.add_argument('--skip-dirty-check', action='store_true',
|
||
|
help='skip checking whether the build is dirty')
|
||
|
parser.add_argument('--verbose', action='store_true',
|
||
|
help='print more diagnostic info')
|
||
|
|
||
|
args, _extras = parser.parse_known_args()
|
||
|
|
||
|
if not os.path.isdir(args.out_dir):
|
||
|
parser.error('OUT_DIR "%s" does not exist.' % args.out_dir)
|
||
|
|
||
|
if not args.skip_dirty_check and not IsBuildClean(args.out_dir):
|
||
|
dirty_msg = 'OUT_DIR looks dirty. You need to build all there.'
|
||
|
if args.json:
|
||
|
# Assume running on the bots. Silently skip this step.
|
||
|
# This is possible because "analyze" step can be wrong due to
|
||
|
# underspecified header files. See crbug.com/725877
|
||
|
print dirty_msg
|
||
|
DumpJson([])
|
||
|
return 0
|
||
|
else:
|
||
|
# Assume running interactively.
|
||
|
parser.error(dirty_msg)
|
||
|
|
||
|
d_q = Queue()
|
||
|
d_p = Process(target=GetHeadersFromNinja, args=(args.out_dir, True, d_q,))
|
||
|
d_p.start()
|
||
|
|
||
|
gn_q = Queue()
|
||
|
gn_p = Process(target=GetHeadersFromGN, args=(args.out_dir, gn_q,))
|
||
|
gn_p.start()
|
||
|
|
||
|
deps_q = Queue()
|
||
|
deps_p = Process(target=GetDepsPrefixes, args=(deps_q,))
|
||
|
deps_p.start()
|
||
|
|
||
|
d, d_err = d_q.get()
|
||
|
gn, gn_err = gn_q.get()
|
||
|
missing = set(d.keys()) - gn
|
||
|
nonexisting = GetNonExistingFiles(gn)
|
||
|
|
||
|
deps, deps_err = deps_q.get()
|
||
|
missing = FilterOutDepsedRepo(missing, deps)
|
||
|
nonexisting = FilterOutDepsedRepo(nonexisting, deps)
|
||
|
|
||
|
d_p.join()
|
||
|
gn_p.join()
|
||
|
deps_p.join()
|
||
|
|
||
|
if d_err:
|
||
|
PrintError(d_err)
|
||
|
if gn_err:
|
||
|
PrintError(gn_err)
|
||
|
if deps_err:
|
||
|
PrintError(deps_err)
|
||
|
if len(GetNonExistingFiles(d)) > 0:
|
||
|
print 'Non-existing files in ninja deps:', GetNonExistingFiles(d)
|
||
|
PrintError('Found non-existing files in ninja deps. You should ' +
|
||
|
'build all in OUT_DIR.')
|
||
|
if len(d) == 0:
|
||
|
PrintError('OUT_DIR looks empty. You should build all there.')
|
||
|
if any((('/gen/' in i) for i in nonexisting)):
|
||
|
PrintError('OUT_DIR looks wrong. You should build all there.')
|
||
|
|
||
|
if args.whitelist:
|
||
|
whitelist = ParseWhiteList(open(args.whitelist).read())
|
||
|
missing -= whitelist
|
||
|
nonexisting -= whitelist
|
||
|
|
||
|
missing = sorted(missing)
|
||
|
nonexisting = sorted(nonexisting)
|
||
|
|
||
|
DumpJson(sorted(missing + nonexisting))
|
||
|
|
||
|
if len(missing) == 0 and len(nonexisting) == 0:
|
||
|
return 0
|
||
|
|
||
|
if len(missing) > 0:
|
||
|
print '\nThe following files should be included in gn files:'
|
||
|
for i in missing:
|
||
|
print i
|
||
|
|
||
|
if len(nonexisting) > 0:
|
||
|
print '\nThe following non-existing files should be removed from gn files:'
|
||
|
for i in nonexisting:
|
||
|
print i
|
||
|
|
||
|
if args.verbose:
|
||
|
# Only get detailed obj dependency here since it is slower.
|
||
|
GetHeadersFromNinja(args.out_dir, False, d_q)
|
||
|
d, d_err = d_q.get()
|
||
|
print '\nDetailed dependency info:'
|
||
|
for f in missing:
|
||
|
print f
|
||
|
for cc in d[f]:
|
||
|
print ' ', cc
|
||
|
|
||
|
print '\nMissing headers sorted by number of affected object files:'
|
||
|
count = {k: len(v) for (k, v) in d.iteritems()}
|
||
|
for f in sorted(count, key=count.get, reverse=True):
|
||
|
if f in missing:
|
||
|
print count[f], f
|
||
|
|
||
|
return 1
|
||
|
|
||
|
|
||
|
if __name__ == '__main__':
|
||
|
sys.exit(main())
|