1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
tools / download_optimization_profile.py [blame]
#!/usr/bin/env vpython3
# Copyright 2018 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""This script is used to update local profiles (AFDO, PGO or orderfiles)
This uses profiles of Chrome, or orderfiles for compiling or linking. Though the
profiles are available externally, the bucket they sit in is otherwise
unreadable by non-Googlers. Gsutil usage with this bucket is therefore quite
awkward: you can't do anything but `cp` certain files with an external account,
and you can't even do that if you're not yet authenticated.
No authentication is necessary if you pull these profiles directly over https.
"""
from __future__ import print_function
import argparse
import contextlib
import os
import subprocess
import sys
from urllib.request import urlopen
GS_HTTP_URL = 'https://storage.googleapis.com'
def ReadUpToDateProfileName(newest_profile_name_path):
with open(newest_profile_name_path) as f:
return f.read().strip()
def ReadLocalProfileName(local_profile_name_path):
try:
with open(local_profile_name_path) as f:
return f.read().strip()
except IOError:
# Assume it either didn't exist, or we couldn't read it. In either case, we
# should probably grab a new profile (and, in doing so, make this file sane
# again)
return None
def WriteLocalProfileName(name, local_profile_name_path):
with open(local_profile_name_path, 'w') as f:
f.write(name)
def CheckCallOrExit(cmd):
proc = subprocess.Popen(cmd,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding='utf-8')
stdout, stderr = proc.communicate()
exit_code = proc.wait()
if not exit_code:
return
complaint_lines = [
'## %s failed with exit code %d' % (cmd[0], exit_code),
'## Full command: %s' % cmd,
'## Stdout:\n' + stdout,
'## Stderr:\n' + stderr,
]
print('\n'.join(complaint_lines), file=sys.stderr)
sys.exit(1)
def RetrieveProfile(desired_profile_name, out_path, gs_url_base):
# vpython is > python 2.7.9, so we can expect urllib to validate HTTPS certs
# properly.
ext = os.path.splitext(desired_profile_name)[1]
if ext in ['.bz2', '.xz']:
# For extension that requires explicit decompression, decompression will
# change the eventual file names by dropping the extension, and that's why
# an extra extension is appended here to make sure that the decompressed
# file path matches the |out_path| passed in as parameter.
out_path += ext
gs_prefix = 'gs://'
if not desired_profile_name.startswith(gs_prefix):
gs_url = '/'.join([GS_HTTP_URL, gs_url_base, desired_profile_name])
else:
gs_url = '/'.join([GS_HTTP_URL, desired_profile_name[len(gs_prefix):]])
with contextlib.closing(urlopen(gs_url)) as u:
with open(out_path, 'wb') as f:
while True:
buf = u.read(4096)
if not buf:
break
f.write(buf)
if ext == '.bz2':
# NOTE: we can't use Python's bzip module, since it doesn't support
# multi-stream bzip files. It will silently succeed and give us a garbage
# profile.
# bzip2 removes the compressed file on success.
CheckCallOrExit(['bzip2', '-d', out_path])
elif ext == '.xz':
# ...And we can't use the `lzma` module, since it was introduced in python3.
# xz removes the compressed file on success.
CheckCallOrExit(['xz', '-d', out_path])
def main():
parser = argparse.ArgumentParser(
description=__doc__, formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument(
'--newest_state',
required=True,
help='Path to the file with name of the newest profile. '
'We use this file to track the name of the newest profile '
'we should pull.')
parser.add_argument(
'--local_state',
required=True,
help='Path of the file storing name of the local profile. '
'We use this file to track the most recent profile we\'ve '
'successfully pulled.')
parser.add_argument(
'--gs_url_base',
required=True,
help='The base GS URL to search for the profile.')
parser.add_argument(
'--output_name',
required=True,
help='Output name of the downloaded and uncompressed profile.')
parser.add_argument(
'-f',
'--force',
action='store_true',
help='Fetch a profile even if the local one is current.')
args = parser.parse_args()
up_to_date_profile = ReadUpToDateProfileName(args.newest_state)
if not args.force:
local_profile_name = ReadLocalProfileName(args.local_state)
# In a perfect world, the local profile should always exist if we
# successfully read local_profile_name. If it's gone, though, the user
# probably removed it as a way to get us to download it again.
if local_profile_name == up_to_date_profile \
and os.path.exists(args.output_name):
return 0
new_tmpfile = args.output_name + '.new'
RetrieveProfile(up_to_date_profile, new_tmpfile, args.gs_url_base)
os.rename(new_tmpfile, args.output_name)
WriteLocalProfileName(up_to_date_profile, args.local_state)
if __name__ == '__main__':
sys.exit(main())