# Copyright (c) 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Bootstrap Chrome Telemetry by downloading all its files from SVN servers. Requires a DEPS file to specify which directories on which SVN servers are required to run Telemetry. Format of that DEPS file is a subset of the normal DEPS file format[1]; currently only only the "deps" dictionary is supported and nothing else. Fetches all files in the specified directories using WebDAV (SVN is WebDAV under the hood). [1] http://dev.chromium.org/developers/how-tos/depottools#TOC-DEPS-file """ import imp import logging import os import urllib import urlparse # Link to file containing the 'davclient' WebDAV client library. _DAVCLIENT_URL = ('https://src.chromium.org/viewvc/chrome/trunk/src/tools/' + 'telemetry/third_party/davclient/davclient.py') # Dummy module for Davclient. _davclient = None def _download_and_import_davclient_module(): """Dynamically import davclient helper library.""" global _davclient davclient_src = urllib.urlopen(_DAVCLIENT_URL).read() _davclient = imp.new_module('davclient') exec davclient_src in _davclient.__dict__ class DAVClientWrapper(): """Knows how to retrieve subdirectories and files from WebDAV/SVN servers.""" def __init__(self, root_url): """Initialize SVN server root_url, save files to local dest_dir. Args: root_url: string url of SVN/WebDAV server """ self.root_url = root_url self.client = _davclient.DAVClient(root_url) @staticmethod def __norm_path_keys(dict_with_path_keys): """Returns a dictionary with os.path.normpath called on every key.""" return dict((os.path.normpath(k), v) for (k, v) in dict_with_path_keys.items()) def GetDirList(self, path): """Returns string names of all files and subdirs of path on the server.""" props = self.__norm_path_keys(self.client.propfind(path, depth=1)) # remove this path del props[os.path.normpath(path)] return [os.path.basename(p) for p in props.keys()] def IsFile(self, path): """Returns True if the path is a file on the server, False if directory.""" props = self.__norm_path_keys(self.client.propfind(path, depth=1)) return props[os.path.normpath(path)]['resourcetype'] is None def Traverse(self, src_path, dst_path): """Walks the directory hierarchy pointed to by src_path download all files. Recursively walks src_path and saves all files and subfolders into dst_path. Args: src_path: string path on SVN server to save (absolute path on server). dest_path: string local path (relative or absolute) to save to. """ if self.IsFile(src_path): if not os.path.exists(os.path.dirname(dst_path)): logging.info("creating %s", os.path.dirname(dst_path)) os.makedirs(os.path.dirname(dst_path)) logging.info("Saving %s to %s", self.root_url + src_path, dst_path) urllib.urlretrieve(self.root_url + src_path, dst_path) return else: for subdir in self.GetDirList(src_path): self.Traverse(os.path.join(src_path, subdir), os.path.join(dst_path, subdir)) def ListAllDepsPaths(deps_content): """Recursively returns a list of all paths indicated in this deps file. Note that this discards information about where path dependencies come from, so this is only useful in the context of a Chromium source checkout that has already fetched all dependencies. Args: deps_content: String containing deps information to be evaluated, in the format given in the header of this file. Returns: A list of string paths starting under src that are required by the given deps file, and all of its sub-dependencies. This amounts to the keys of the 'deps' dictionary. """ chrome_root = os.path.dirname(__file__) while os.path.basename(chrome_root) != 'src': chrome_root = os.path.abspath(os.path.join(chrome_root, '..')) deps = imp.new_module('deps') exec deps_content in deps.__dict__ deps_paths = deps.deps.keys() if hasattr(deps, 'deps_includes'): for path in deps.deps_includes.keys(): # Need to localize the paths. path = os.path.join(chrome_root, '..', path) deps_paths = deps_paths + ListAllDepsPaths(open(path).read()) return deps_paths def DownloadDepsURL(destination_dir, url): """Wrapper around DownloadDeps that takes a string URL to the deps file. Args: destination_dir: String path to local directory to download files into. url: URL of deps file (see DownloadDeps for format). """ logging.warning('Downloading deps from %s...', url) DownloadDeps(destination_dir, urllib.urlopen(url).read()) def DownloadDeps(destination_dir, deps_content): """Saves all the dependencies in deps_path. Reads deps_content, assuming the contents are in the simple DEPS-like file format specified in the header of this file, then download all files/directories listed to the destination_dir. Args: destination_dir: String path to directory to download files into. deps_content: String containing deps information to be evaluated. """ # TODO(wiltzius): Add a parameter for which revision to pull. _download_and_import_davclient_module() deps = imp.new_module('deps') exec deps_content in deps.__dict__ for dst_path, src_path in deps.deps.iteritems(): full_dst_path = os.path.join(destination_dir, dst_path) parsed_url = urlparse.urlparse(src_path) root_url = parsed_url.scheme + '://' + parsed_url.netloc dav_client = DAVClientWrapper(root_url) dav_client.Traverse(parsed_url.path, full_dst_path) if hasattr(deps, 'deps_includes'): for url in deps.deps_includes.values(): DownloadDepsURL(destination_dir, url)