163 lines
5.8 KiB
Python
163 lines
5.8 KiB
Python
|
# Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
||
|
# Use of this source code is governed by a BSD-style license that can be
|
||
|
# found in the LICENSE file.
|
||
|
|
||
|
"""Bootstrap Chrome Telemetry by downloading all its files from SVN servers.
|
||
|
|
||
|
Requires a DEPS file to specify which directories on which SVN servers
|
||
|
are required to run Telemetry. Format of that DEPS file is a subset of the
|
||
|
normal DEPS file format[1]; currently only only the "deps" dictionary is
|
||
|
supported and nothing else.
|
||
|
|
||
|
Fetches all files in the specified directories using WebDAV (SVN is WebDAV under
|
||
|
the hood).
|
||
|
|
||
|
[1] http://dev.chromium.org/developers/how-tos/depottools#TOC-DEPS-file
|
||
|
"""
|
||
|
|
||
|
import imp
|
||
|
import logging
|
||
|
import os
|
||
|
import urllib
|
||
|
import urlparse
|
||
|
|
||
|
# Link to file containing the 'davclient' WebDAV client library.
|
||
|
_DAVCLIENT_URL = ('https://src.chromium.org/viewvc/chrome/trunk/src/tools/' +
|
||
|
'telemetry/third_party/davclient/davclient.py')
|
||
|
|
||
|
# Dummy module for Davclient.
|
||
|
_davclient = None
|
||
|
|
||
|
def _download_and_import_davclient_module():
|
||
|
"""Dynamically import davclient helper library."""
|
||
|
global _davclient
|
||
|
davclient_src = urllib.urlopen(_DAVCLIENT_URL).read()
|
||
|
_davclient = imp.new_module('davclient')
|
||
|
exec davclient_src in _davclient.__dict__
|
||
|
|
||
|
|
||
|
class DAVClientWrapper():
|
||
|
"""Knows how to retrieve subdirectories and files from WebDAV/SVN servers."""
|
||
|
|
||
|
def __init__(self, root_url):
|
||
|
"""Initialize SVN server root_url, save files to local dest_dir.
|
||
|
|
||
|
Args:
|
||
|
root_url: string url of SVN/WebDAV server
|
||
|
"""
|
||
|
self.root_url = root_url
|
||
|
self.client = _davclient.DAVClient(root_url)
|
||
|
|
||
|
@staticmethod
|
||
|
def __norm_path_keys(dict_with_path_keys):
|
||
|
"""Returns a dictionary with os.path.normpath called on every key."""
|
||
|
return dict((os.path.normpath(k), v) for (k, v) in
|
||
|
dict_with_path_keys.items())
|
||
|
|
||
|
def GetDirList(self, path):
|
||
|
"""Returns string names of all files and subdirs of path on the server."""
|
||
|
props = self.__norm_path_keys(self.client.propfind(path, depth=1))
|
||
|
# remove this path
|
||
|
del props[os.path.normpath(path)]
|
||
|
return [os.path.basename(p) for p in props.keys()]
|
||
|
|
||
|
def IsFile(self, path):
|
||
|
"""Returns True if the path is a file on the server, False if directory."""
|
||
|
props = self.__norm_path_keys(self.client.propfind(path, depth=1))
|
||
|
return props[os.path.normpath(path)]['resourcetype'] is None
|
||
|
|
||
|
def Traverse(self, src_path, dst_path):
|
||
|
"""Walks the directory hierarchy pointed to by src_path download all files.
|
||
|
|
||
|
Recursively walks src_path and saves all files and subfolders into
|
||
|
dst_path.
|
||
|
|
||
|
Args:
|
||
|
src_path: string path on SVN server to save (absolute path on server).
|
||
|
dest_path: string local path (relative or absolute) to save to.
|
||
|
"""
|
||
|
if self.IsFile(src_path):
|
||
|
if not os.path.exists(os.path.dirname(dst_path)):
|
||
|
logging.info("creating %s", os.path.dirname(dst_path))
|
||
|
os.makedirs(os.path.dirname(dst_path))
|
||
|
logging.info("Saving %s to %s", self.root_url + src_path, dst_path)
|
||
|
urllib.urlretrieve(self.root_url + src_path, dst_path)
|
||
|
return
|
||
|
else:
|
||
|
for subdir in self.GetDirList(src_path):
|
||
|
self.Traverse(os.path.join(src_path, subdir),
|
||
|
os.path.join(dst_path, subdir))
|
||
|
|
||
|
|
||
|
def ListAllDepsPaths(deps_content):
|
||
|
"""Recursively returns a list of all paths indicated in this deps file.
|
||
|
|
||
|
Note that this discards information about where path dependencies come from,
|
||
|
so this is only useful in the context of a Chromium source checkout that has
|
||
|
already fetched all dependencies.
|
||
|
|
||
|
Args:
|
||
|
deps_content: String containing deps information to be evaluated, in the
|
||
|
format given in the header of this file.
|
||
|
Returns: A list of string paths starting under src that are required by the
|
||
|
given deps file, and all of its sub-dependencies. This amounts to
|
||
|
the keys of the 'deps' dictionary.
|
||
|
"""
|
||
|
chrome_root = os.path.dirname(__file__)
|
||
|
while os.path.basename(chrome_root) != 'src':
|
||
|
chrome_root = os.path.abspath(os.path.join(chrome_root, '..'))
|
||
|
deps = imp.new_module('deps')
|
||
|
exec deps_content in deps.__dict__
|
||
|
|
||
|
deps_paths = deps.deps.keys()
|
||
|
|
||
|
if hasattr(deps, 'deps_includes'):
|
||
|
for path in deps.deps_includes.keys():
|
||
|
# Need to localize the paths.
|
||
|
path = os.path.join(chrome_root, '..', path)
|
||
|
deps_paths = deps_paths + ListAllDepsPaths(open(path).read())
|
||
|
|
||
|
return deps_paths
|
||
|
|
||
|
|
||
|
def DownloadDepsURL(destination_dir, url):
|
||
|
"""Wrapper around DownloadDeps that takes a string URL to the deps file.
|
||
|
|
||
|
Args:
|
||
|
destination_dir: String path to local directory to download files into.
|
||
|
url: URL of deps file (see DownloadDeps for format).
|
||
|
"""
|
||
|
logging.warning('Downloading deps from %s...', url)
|
||
|
DownloadDeps(destination_dir, urllib.urlopen(url).read())
|
||
|
|
||
|
|
||
|
def DownloadDeps(destination_dir, deps_content):
|
||
|
"""Saves all the dependencies in deps_path.
|
||
|
|
||
|
Reads deps_content, assuming the contents are in the simple DEPS-like file
|
||
|
format specified in the header of this file, then download all
|
||
|
files/directories listed to the destination_dir.
|
||
|
|
||
|
Args:
|
||
|
destination_dir: String path to directory to download files into.
|
||
|
deps_content: String containing deps information to be evaluated.
|
||
|
"""
|
||
|
# TODO(wiltzius): Add a parameter for which revision to pull.
|
||
|
_download_and_import_davclient_module()
|
||
|
|
||
|
deps = imp.new_module('deps')
|
||
|
exec deps_content in deps.__dict__
|
||
|
|
||
|
for dst_path, src_path in deps.deps.iteritems():
|
||
|
full_dst_path = os.path.join(destination_dir, dst_path)
|
||
|
parsed_url = urlparse.urlparse(src_path)
|
||
|
root_url = parsed_url.scheme + '://' + parsed_url.netloc
|
||
|
|
||
|
dav_client = DAVClientWrapper(root_url)
|
||
|
dav_client.Traverse(parsed_url.path, full_dst_path)
|
||
|
|
||
|
if hasattr(deps, 'deps_includes'):
|
||
|
for url in deps.deps_includes.values():
|
||
|
DownloadDepsURL(destination_dir, url)
|
||
|
|