| # Copyright (c) 2006-2009 The Chromium Authors. All rights reserved. |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| |
| """SCM-specific utility classes.""" |
| |
| import cStringIO |
| import glob |
| import os |
| import re |
| import shutil |
| import subprocess |
| import sys |
| import tempfile |
| import time |
| import xml.dom.minidom |
| |
| import gclient_utils |
| |
| def ValidateEmail(email): |
| return (re.match(r"^[a-zA-Z0-9._%-+]+@[a-zA-Z0-9._%-]+.[a-zA-Z]{2,6}$", email) |
| is not None) |
| |
| |
| def GetCasedPath(path): |
| """Elcheapos way to get the real path case on Windows.""" |
| if sys.platform.startswith('win') and os.path.exists(path): |
| # Reconstruct the path. |
| path = os.path.abspath(path) |
| paths = path.split('\\') |
| for i in range(len(paths)): |
| if i == 0: |
| # Skip drive letter. |
| continue |
| subpath = '\\'.join(paths[:i+1]) |
| prev = len('\\'.join(paths[:i])) |
| # glob.glob will return the cased path for the last item only. This is why |
| # we are calling it in a loop. Extract the data we want and put it back |
| # into the list. |
| paths[i] = glob.glob(subpath + '*')[0][prev+1:len(subpath)] |
| path = '\\'.join(paths) |
| return path |
| |
| |
| def GenFakeDiff(filename): |
| """Generates a fake diff from a file.""" |
| file_content = gclient_utils.FileRead(filename, 'rb').splitlines(True) |
| filename = filename.replace(os.sep, '/') |
| nb_lines = len(file_content) |
| # We need to use / since patch on unix will fail otherwise. |
| data = cStringIO.StringIO() |
| data.write("Index: %s\n" % filename) |
| data.write('=' * 67 + '\n') |
| # Note: Should we use /dev/null instead? |
| data.write("--- %s\n" % filename) |
| data.write("+++ %s\n" % filename) |
| data.write("@@ -0,0 +1,%d @@\n" % nb_lines) |
| # Prepend '+' to every lines. |
| for line in file_content: |
| data.write('+') |
| data.write(line) |
| result = data.getvalue() |
| data.close() |
| return result |
| |
| |
| class GIT(object): |
| @staticmethod |
| def Capture(args, in_directory=None, print_error=True, error_ok=False): |
| """Runs git, capturing output sent to stdout as a string. |
| |
| Args: |
| args: A sequence of command line parameters to be passed to git. |
| in_directory: The directory where git is to be run. |
| |
| Returns: |
| The output sent to stdout as a string. |
| """ |
| try: |
| return gclient_utils.CheckCall(['git'] + args, in_directory, print_error) |
| except gclient_utils.CheckCallError: |
| if error_ok: |
| return ('', '') |
| raise |
| |
| @staticmethod |
| def CaptureStatus(files, upstream_branch=None): |
| """Returns git status. |
| |
| @files can be a string (one file) or a list of files. |
| |
| Returns an array of (status, file) tuples.""" |
| if upstream_branch is None: |
| upstream_branch = GIT.GetUpstreamBranch(os.getcwd()) |
| if upstream_branch is None: |
| raise Exception("Cannot determine upstream branch") |
| command = ["diff", "--name-status", "-r", "%s..." % upstream_branch] |
| if not files: |
| pass |
| elif isinstance(files, basestring): |
| command.append(files) |
| else: |
| command.extend(files) |
| |
| status = GIT.Capture(command)[0].rstrip() |
| results = [] |
| if status: |
| for statusline in status.split('\n'): |
| m = re.match('^(\w)\t(.+)$', statusline) |
| if not m: |
| raise Exception("status currently unsupported: %s" % statusline) |
| results.append(('%s ' % m.group(1), m.group(2))) |
| return results |
| |
| @staticmethod |
| def GetEmail(repo_root): |
| """Retrieves the user email address if known.""" |
| # We could want to look at the svn cred when it has a svn remote but it |
| # should be fine for now, users should simply configure their git settings. |
| return GIT.Capture(['config', 'user.email'], |
| repo_root, error_ok=True)[0].strip() |
| |
| @staticmethod |
| def ShortBranchName(branch): |
| """Converts a name like 'refs/heads/foo' to just 'foo'.""" |
| return branch.replace('refs/heads/', '') |
| |
| @staticmethod |
| def GetBranchRef(cwd): |
| """Returns the full branch reference, e.g. 'refs/heads/master'.""" |
| return GIT.Capture(['symbolic-ref', 'HEAD'], cwd)[0].strip() |
| |
| @staticmethod |
| def GetBranch(cwd): |
| """Returns the short branch name, e.g. 'master'.""" |
| return GIT.ShortBranchName(GIT.GetBranchRef(cwd)) |
| |
| @staticmethod |
| def IsGitSvn(cwd): |
| """Returns true if this repo looks like it's using git-svn.""" |
| # If you have any "svn-remote.*" config keys, we think you're using svn. |
| try: |
| GIT.Capture(['config', '--get-regexp', r'^svn-remote\.'], cwd) |
| return True |
| except gclient_utils.CheckCallError: |
| return False |
| |
| @staticmethod |
| def GetSVNBranch(cwd): |
| """Returns the svn branch name if found.""" |
| # Try to figure out which remote branch we're based on. |
| # Strategy: |
| # 1) find all git-svn branches and note their svn URLs. |
| # 2) iterate through our branch history and match up the URLs. |
| |
| # regexp matching the git-svn line that contains the URL. |
| git_svn_re = re.compile(r'^\s*git-svn-id: (\S+)@', re.MULTILINE) |
| |
| # Get the refname and svn url for all refs/remotes/*. |
| remotes = GIT.Capture( |
| ['for-each-ref', '--format=%(refname)', 'refs/remotes'], |
| cwd)[0].splitlines() |
| svn_refs = {} |
| for ref in remotes: |
| match = git_svn_re.search( |
| GIT.Capture(['cat-file', '-p', ref], cwd)[0]) |
| # Prefer origin/HEAD over all others. |
| if match and (match.group(1) not in svn_refs or |
| ref == "refs/remotes/origin/HEAD"): |
| svn_refs[match.group(1)] = ref |
| |
| svn_branch = '' |
| if len(svn_refs) == 1: |
| # Only one svn branch exists -- seems like a good candidate. |
| svn_branch = svn_refs.values()[0] |
| elif len(svn_refs) > 1: |
| # We have more than one remote branch available. We don't |
| # want to go through all of history, so read a line from the |
| # pipe at a time. |
| # The -100 is an arbitrary limit so we don't search forever. |
| cmd = ['git', 'log', '-100', '--pretty=medium'] |
| proc = gclient_utils.Popen(cmd, stdout=subprocess.PIPE, cwd=cwd) |
| for line in proc.stdout: |
| match = git_svn_re.match(line) |
| if match: |
| url = match.group(1) |
| if url in svn_refs: |
| svn_branch = svn_refs[url] |
| proc.stdout.close() # Cut pipe. |
| break |
| return svn_branch |
| |
| @staticmethod |
| def FetchUpstreamTuple(cwd): |
| """Returns a tuple containg remote and remote ref, |
| e.g. 'origin', 'refs/heads/master' |
| Tries to be intelligent and understand git-svn. |
| """ |
| remote = '.' |
| branch = GIT.GetBranch(cwd) |
| upstream_branch = None |
| upstream_branch = GIT.Capture( |
| ['config', 'branch.%s.merge' % branch], in_directory=cwd, |
| error_ok=True)[0].strip() |
| if upstream_branch: |
| remote = GIT.Capture( |
| ['config', 'branch.%s.remote' % branch], |
| in_directory=cwd, error_ok=True)[0].strip() |
| else: |
| # Fall back on trying a git-svn upstream branch. |
| if GIT.IsGitSvn(cwd): |
| upstream_branch = GIT.GetSVNBranch(cwd) |
| else: |
| # Else, try to guess the origin remote. |
| remote_branches = GIT.Capture( |
| ['branch', '-r'], in_directory=cwd)[0].split() |
| if 'origin/master' in remote_branches: |
| # Fall back on origin/master if it exits. |
| remote = 'origin' |
| upstream_branch = 'refs/heads/master' |
| elif 'origin/trunk' in remote_branches: |
| # Fall back on origin/trunk if it exists. Generally a shared |
| # git-svn clone |
| remote = 'origin' |
| upstream_branch = 'refs/heads/trunk' |
| else: |
| # Give up. |
| remote = None |
| upstream_branch = None |
| return remote, upstream_branch |
| |
| @staticmethod |
| def GetUpstreamBranch(cwd): |
| """Gets the current branch's upstream branch.""" |
| remote, upstream_branch = GIT.FetchUpstreamTuple(cwd) |
| if remote != '.' and upstream_branch: |
| upstream_branch = upstream_branch.replace('heads', 'remotes/' + remote) |
| return upstream_branch |
| |
| @staticmethod |
| def GenerateDiff(cwd, branch=None, branch_head='HEAD', full_move=False, |
| files=None): |
| """Diffs against the upstream branch or optionally another branch. |
| |
| full_move means that move or copy operations should completely recreate the |
| files, usually in the prospect to apply the patch for a try job.""" |
| if not branch: |
| branch = GIT.GetUpstreamBranch(cwd) |
| command = ['diff', '-p', '--no-prefix', '--no-ext-diff', |
| branch + "..." + branch_head] |
| if not full_move: |
| command.append('-C') |
| # TODO(maruel): --binary support. |
| if files: |
| command.append('--') |
| command.extend(files) |
| diff = GIT.Capture(command, cwd)[0].splitlines(True) |
| for i in range(len(diff)): |
| # In the case of added files, replace /dev/null with the path to the |
| # file being added. |
| if diff[i].startswith('--- /dev/null'): |
| diff[i] = '--- %s' % diff[i+1][4:] |
| return ''.join(diff) |
| |
| @staticmethod |
| def GetDifferentFiles(cwd, branch=None, branch_head='HEAD'): |
| """Returns the list of modified files between two branches.""" |
| if not branch: |
| branch = GIT.GetUpstreamBranch(cwd) |
| command = ['diff', '--name-only', branch + "..." + branch_head] |
| return GIT.Capture(command, cwd)[0].splitlines(False) |
| |
| @staticmethod |
| def GetPatchName(cwd): |
| """Constructs a name for this patch.""" |
| short_sha = GIT.Capture(['rev-parse', '--short=4', 'HEAD'], cwd)[0].strip() |
| return "%s#%s" % (GIT.GetBranch(cwd), short_sha) |
| |
| @staticmethod |
| def GetCheckoutRoot(path): |
| """Returns the top level directory of a git checkout as an absolute path. |
| """ |
| root = GIT.Capture(['rev-parse', '--show-cdup'], path)[0].strip() |
| return os.path.abspath(os.path.join(path, root)) |
| |
| @staticmethod |
| def AssertVersion(min_version): |
| """Asserts git's version is at least min_version.""" |
| def only_int(val): |
| if val.isdigit(): |
| return int(val) |
| else: |
| return 0 |
| current_version = GIT.Capture(['--version'])[0].split()[-1] |
| current_version_list = map(only_int, current_version.split('.')) |
| for min_ver in map(int, min_version.split('.')): |
| ver = current_version_list.pop(0) |
| if ver < min_ver: |
| return (False, current_version) |
| elif ver > min_ver: |
| return (True, current_version) |
| return (True, current_version) |
| |
| |
| class SVN(object): |
| current_version = None |
| |
| @staticmethod |
| def Capture(args, in_directory=None, print_error=True): |
| """Runs svn, capturing output sent to stdout as a string. |
| |
| Args: |
| args: A sequence of command line parameters to be passed to svn. |
| in_directory: The directory where svn is to be run. |
| |
| Returns: |
| The output sent to stdout as a string. |
| """ |
| stderr = None |
| if not print_error: |
| stderr = subprocess.PIPE |
| return gclient_utils.Popen(['svn'] + args, cwd=in_directory, |
| stdout=subprocess.PIPE, stderr=stderr).communicate()[0] |
| |
| @staticmethod |
| def RunAndGetFileList(verbose, args, cwd, file_list, stdout=None): |
| """Runs svn checkout, update, or status, output to stdout. |
| |
| The first item in args must be either "checkout", "update", or "status". |
| |
| svn's stdout is parsed to collect a list of files checked out or updated. |
| These files are appended to file_list. svn's stdout is also printed to |
| sys.stdout as in Run. |
| |
| Args: |
| verbose: If True, uses verbose output |
| args: A sequence of command line parameters to be passed to svn. |
| cwd: The directory where svn is to be run. |
| |
| Raises: |
| Error: An error occurred while running the svn command. |
| """ |
| stdout = stdout or sys.stdout |
| |
| # svn update and svn checkout use the same pattern: the first three columns |
| # are for file status, property status, and lock status. This is followed |
| # by two spaces, and then the path to the file. |
| update_pattern = '^... (.*)$' |
| |
| # The first three columns of svn status are the same as for svn update and |
| # svn checkout. The next three columns indicate addition-with-history, |
| # switch, and remote lock status. This is followed by one space, and then |
| # the path to the file. |
| status_pattern = '^...... (.*)$' |
| |
| # args[0] must be a supported command. This will blow up if it's something |
| # else, which is good. Note that the patterns are only effective when |
| # these commands are used in their ordinary forms, the patterns are invalid |
| # for "svn status --show-updates", for example. |
| pattern = { |
| 'checkout': update_pattern, |
| 'status': status_pattern, |
| 'update': update_pattern, |
| }[args[0]] |
| compiled_pattern = re.compile(pattern) |
| # Place an upper limit. |
| backoff_time = 5 |
| for _ in range(10): |
| previous_list_len = len(file_list) |
| failure = [] |
| |
| def CaptureMatchingLines(line): |
| match = compiled_pattern.search(line) |
| if match: |
| file_list.append(match.group(1)) |
| if line.startswith('svn: '): |
| failure.append(line) |
| |
| try: |
| gclient_utils.CheckCallAndFilterAndHeader( |
| ['svn'] + args, |
| cwd=cwd, |
| always=verbose, |
| filter_fn=CaptureMatchingLines, |
| stdout=stdout) |
| except gclient_utils.Error: |
| def IsKnownFailure(): |
| for x in failure: |
| if (x.startswith('svn: OPTIONS of') or |
| x.startswith('svn: PROPFIND of') or |
| x.startswith('svn: REPORT of') or |
| x.startswith('svn: Unknown hostname') or |
| x.startswith('svn: Server sent unexpected return value')): |
| return True |
| return False |
| |
| # Subversion client is really misbehaving with Google Code. |
| if args[0] == 'checkout': |
| # Ensure at least one file was checked out, otherwise *delete* the |
| # directory. |
| if len(file_list) == previous_list_len: |
| if not IsKnownFailure(): |
| # No known svn error was found, bail out. |
| raise |
| # No file were checked out, so make sure the directory is |
| # deleted in case it's messed up and try again. |
| # Warning: It's bad, it assumes args[2] is the directory |
| # argument. |
| if os.path.isdir(args[2]): |
| gclient_utils.RemoveDirectory(args[2]) |
| else: |
| # Progress was made, convert to update since an aborted checkout |
| # is now an update. |
| args = ['update'] + args[1:] |
| else: |
| # It was an update or export. |
| # We enforce that some progress has been made or a known failure. |
| if len(file_list) == previous_list_len and not IsKnownFailure(): |
| # No known svn error was found and no progress, bail out. |
| raise |
| print "Sleeping %.1f seconds and retrying...." % backoff_time |
| time.sleep(backoff_time) |
| backoff_time *= 1.3 |
| continue |
| break |
| |
| @staticmethod |
| def CaptureInfo(relpath, in_directory=None, print_error=True): |
| """Returns a dictionary from the svn info output for the given file. |
| |
| Args: |
| relpath: The directory where the working copy resides relative to |
| the directory given by in_directory. |
| in_directory: The directory where svn is to be run. |
| """ |
| output = SVN.Capture(["info", "--xml", relpath], in_directory, print_error) |
| dom = gclient_utils.ParseXML(output) |
| result = {} |
| if dom: |
| GetNamedNodeText = gclient_utils.GetNamedNodeText |
| GetNodeNamedAttributeText = gclient_utils.GetNodeNamedAttributeText |
| def C(item, f): |
| if item is not None: |
| return f(item) |
| # /info/entry/ |
| # url |
| # reposityory/(root|uuid) |
| # wc-info/(schedule|depth) |
| # commit/(author|date) |
| # str() the results because they may be returned as Unicode, which |
| # interferes with the higher layers matching up things in the deps |
| # dictionary. |
| result['Repository Root'] = C(GetNamedNodeText(dom, 'root'), str) |
| result['URL'] = C(GetNamedNodeText(dom, 'url'), str) |
| result['UUID'] = C(GetNamedNodeText(dom, 'uuid'), str) |
| result['Revision'] = C(GetNodeNamedAttributeText(dom, 'entry', |
| 'revision'), |
| int) |
| result['Node Kind'] = C(GetNodeNamedAttributeText(dom, 'entry', 'kind'), |
| str) |
| # Differs across versions. |
| if result['Node Kind'] == 'dir': |
| result['Node Kind'] = 'directory' |
| result['Schedule'] = C(GetNamedNodeText(dom, 'schedule'), str) |
| result['Path'] = C(GetNodeNamedAttributeText(dom, 'entry', 'path'), str) |
| result['Copied From URL'] = C(GetNamedNodeText(dom, 'copy-from-url'), str) |
| result['Copied From Rev'] = C(GetNamedNodeText(dom, 'copy-from-rev'), str) |
| return result |
| |
| @staticmethod |
| def CaptureHeadRevision(url): |
| """Get the head revision of a SVN repository. |
| |
| Returns: |
| Int head revision |
| """ |
| info = SVN.Capture(["info", "--xml", url], os.getcwd()) |
| dom = xml.dom.minidom.parseString(info) |
| return dom.getElementsByTagName('entry')[0].getAttribute('revision') |
| |
| @staticmethod |
| def CaptureBaseRevision(cwd): |
| """Get the base revision of a SVN repository. |
| |
| Returns: |
| Int base revision |
| """ |
| info = SVN.Capture(["info", "--xml"], cwd) |
| dom = xml.dom.minidom.parseString(info) |
| return dom.getElementsByTagName('entry')[0].getAttribute('revision') |
| |
| @staticmethod |
| def CaptureStatus(files): |
| """Returns the svn 1.5 svn status emulated output. |
| |
| @files can be a string (one file) or a list of files. |
| |
| Returns an array of (status, file) tuples.""" |
| command = ["status", "--xml"] |
| if not files: |
| pass |
| elif isinstance(files, basestring): |
| command.append(files) |
| else: |
| command.extend(files) |
| |
| status_letter = { |
| None: ' ', |
| '': ' ', |
| 'added': 'A', |
| 'conflicted': 'C', |
| 'deleted': 'D', |
| 'external': 'X', |
| 'ignored': 'I', |
| 'incomplete': '!', |
| 'merged': 'G', |
| 'missing': '!', |
| 'modified': 'M', |
| 'none': ' ', |
| 'normal': ' ', |
| 'obstructed': '~', |
| 'replaced': 'R', |
| 'unversioned': '?', |
| } |
| dom = gclient_utils.ParseXML(SVN.Capture(command)) |
| results = [] |
| if dom: |
| # /status/target/entry/(wc-status|commit|author|date) |
| for target in dom.getElementsByTagName('target'): |
| #base_path = target.getAttribute('path') |
| for entry in target.getElementsByTagName('entry'): |
| file_path = entry.getAttribute('path') |
| wc_status = entry.getElementsByTagName('wc-status') |
| assert len(wc_status) == 1 |
| # Emulate svn 1.5 status ouput... |
| statuses = [' '] * 7 |
| # Col 0 |
| xml_item_status = wc_status[0].getAttribute('item') |
| if xml_item_status in status_letter: |
| statuses[0] = status_letter[xml_item_status] |
| else: |
| raise Exception('Unknown item status "%s"; please implement me!' % |
| xml_item_status) |
| # Col 1 |
| xml_props_status = wc_status[0].getAttribute('props') |
| if xml_props_status == 'modified': |
| statuses[1] = 'M' |
| elif xml_props_status == 'conflicted': |
| statuses[1] = 'C' |
| elif (not xml_props_status or xml_props_status == 'none' or |
| xml_props_status == 'normal'): |
| pass |
| else: |
| raise Exception('Unknown props status "%s"; please implement me!' % |
| xml_props_status) |
| # Col 2 |
| if wc_status[0].getAttribute('wc-locked') == 'true': |
| statuses[2] = 'L' |
| # Col 3 |
| if wc_status[0].getAttribute('copied') == 'true': |
| statuses[3] = '+' |
| # Col 4 |
| if wc_status[0].getAttribute('switched') == 'true': |
| statuses[4] = 'S' |
| # TODO(maruel): Col 5 and 6 |
| item = (''.join(statuses), file_path) |
| results.append(item) |
| return results |
| |
| @staticmethod |
| def IsMoved(filename): |
| """Determine if a file has been added through svn mv""" |
| return SVN.IsMovedInfo(SVN.CaptureInfo(filename)) |
| |
| @staticmethod |
| def IsMovedInfo(info): |
| """Determine if a file has been added through svn mv""" |
| return (info.get('Copied From URL') and |
| info.get('Copied From Rev') and |
| info.get('Schedule') == 'add') |
| |
| @staticmethod |
| def GetFileProperty(filename, property_name): |
| """Returns the value of an SVN property for the given file. |
| |
| Args: |
| filename: The file to check |
| property_name: The name of the SVN property, e.g. "svn:mime-type" |
| |
| Returns: |
| The value of the property, which will be the empty string if the property |
| is not set on the file. If the file is not under version control, the |
| empty string is also returned. |
| """ |
| output = SVN.Capture(["propget", property_name, filename]) |
| if (output.startswith("svn: ") and |
| output.endswith("is not under version control")): |
| return "" |
| else: |
| return output |
| |
| @staticmethod |
| def DiffItem(filename, full_move=False, revision=None): |
| """Diffs a single file. |
| |
| Should be simple, eh? No it isn't. |
| Be sure to be in the appropriate directory before calling to have the |
| expected relative path. |
| full_move means that move or copy operations should completely recreate the |
| files, usually in the prospect to apply the patch for a try job.""" |
| # If the user specified a custom diff command in their svn config file, |
| # then it'll be used when we do svn diff, which we don't want to happen |
| # since we want the unified diff. Using --diff-cmd=diff doesn't always |
| # work, since they can have another diff executable in their path that |
| # gives different line endings. So we use a bogus temp directory as the |
| # config directory, which gets around these problems. |
| bogus_dir = tempfile.mkdtemp() |
| try: |
| # Use "svn info" output instead of os.path.isdir because the latter fails |
| # when the file is deleted. |
| return SVN._DiffItemInternal(filename, SVN.CaptureInfo(filename), |
| bogus_dir, |
| full_move=full_move, revision=revision) |
| finally: |
| shutil.rmtree(bogus_dir) |
| |
| @staticmethod |
| def _DiffItemInternal(filename, info, bogus_dir, full_move=False, |
| revision=None): |
| """Grabs the diff data.""" |
| command = ["diff", "--config-dir", bogus_dir, filename] |
| if revision: |
| command.extend(['--revision', revision]) |
| data = None |
| if SVN.IsMovedInfo(info): |
| if full_move: |
| if info.get("Node Kind") == "directory": |
| # Things become tricky here. It's a directory copy/move. We need to |
| # diff all the files inside it. |
| # This will put a lot of pressure on the heap. This is why StringIO |
| # is used and converted back into a string at the end. The reason to |
| # return a string instead of a StringIO is that StringIO.write() |
| # doesn't accept a StringIO object. *sigh*. |
| for (dirpath, dirnames, filenames) in os.walk(filename): |
| # Cleanup all files starting with a '.'. |
| for d in dirnames: |
| if d.startswith('.'): |
| dirnames.remove(d) |
| for f in filenames: |
| if f.startswith('.'): |
| filenames.remove(f) |
| for f in filenames: |
| if data is None: |
| data = cStringIO.StringIO() |
| data.write(GenFakeDiff(os.path.join(dirpath, f))) |
| if data: |
| tmp = data.getvalue() |
| data.close() |
| data = tmp |
| else: |
| data = GenFakeDiff(filename) |
| else: |
| if info.get("Node Kind") != "directory": |
| # svn diff on a mv/cp'd file outputs nothing if there was no change. |
| data = SVN.Capture(command, None) |
| if not data: |
| # We put in an empty Index entry so upload.py knows about them. |
| data = "Index: %s\n" % filename.replace(os.sep, '/') |
| # Otherwise silently ignore directories. |
| else: |
| if info.get("Node Kind") != "directory": |
| # Normal simple case. |
| data = SVN.Capture(command, None) |
| # Otherwise silently ignore directories. |
| return data |
| |
| @staticmethod |
| def GenerateDiff(filenames, root=None, full_move=False, revision=None): |
| """Returns a string containing the diff for the given file list. |
| |
| The files in the list should either be absolute paths or relative to the |
| given root. If no root directory is provided, the repository root will be |
| used. |
| The diff will always use relative paths. |
| """ |
| previous_cwd = os.getcwd() |
| root = root or SVN.GetCheckoutRoot(previous_cwd) |
| root = os.path.normcase(os.path.join(root, '')) |
| def RelativePath(path, root): |
| """We must use relative paths.""" |
| if os.path.normcase(path).startswith(root): |
| return path[len(root):] |
| return path |
| # If the user specified a custom diff command in their svn config file, |
| # then it'll be used when we do svn diff, which we don't want to happen |
| # since we want the unified diff. Using --diff-cmd=diff doesn't always |
| # work, since they can have another diff executable in their path that |
| # gives different line endings. So we use a bogus temp directory as the |
| # config directory, which gets around these problems. |
| bogus_dir = tempfile.mkdtemp() |
| try: |
| os.chdir(root) |
| # Cleanup filenames |
| filenames = [RelativePath(f, root) for f in filenames] |
| # Get information about the modified items (files and directories) |
| data = dict([(f, SVN.CaptureInfo(f)) for f in filenames]) |
| diffs = [] |
| if full_move: |
| # Eliminate modified files inside moved/copied directory. |
| for (filename, info) in data.iteritems(): |
| if SVN.IsMovedInfo(info) and info.get("Node Kind") == "directory": |
| # Remove files inside the directory. |
| filenames = [f for f in filenames |
| if not f.startswith(filename + os.path.sep)] |
| for filename in data.keys(): |
| if not filename in filenames: |
| # Remove filtered out items. |
| del data[filename] |
| else: |
| metaheaders = [] |
| for (filename, info) in data.iteritems(): |
| if SVN.IsMovedInfo(info): |
| # for now, the most common case is a head copy, |
| # so let's just encode that as a straight up cp. |
| srcurl = info.get('Copied From URL') |
| root = info.get('Repository Root') |
| rev = int(info.get('Copied From Rev')) |
| assert srcurl.startswith(root) |
| src = srcurl[len(root)+1:] |
| srcinfo = SVN.CaptureInfo(srcurl) |
| if (srcinfo.get('Revision') != rev and |
| SVN.Capture(['diff', '-r', '%d:head' % rev, srcurl])): |
| metaheaders.append("#$ svn cp -r %d %s %s " |
| "### WARNING: note non-trunk copy\n" % |
| (rev, src, filename)) |
| else: |
| metaheaders.append("#$ cp %s %s\n" % (src, |
| filename)) |
| |
| if metaheaders: |
| diffs.append("### BEGIN SVN COPY METADATA\n") |
| diffs.extend(metaheaders) |
| diffs.append("### END SVN COPY METADATA\n") |
| # Now ready to do the actual diff. |
| for filename in sorted(data.iterkeys()): |
| diffs.append(SVN._DiffItemInternal(filename, data[filename], bogus_dir, |
| full_move=full_move, |
| revision=revision)) |
| # Use StringIO since it can be messy when diffing a directory move with |
| # full_move=True. |
| buf = cStringIO.StringIO() |
| for d in filter(None, diffs): |
| buf.write(d) |
| result = buf.getvalue() |
| buf.close() |
| return result |
| finally: |
| os.chdir(previous_cwd) |
| shutil.rmtree(bogus_dir) |
| |
| @staticmethod |
| def GetEmail(repo_root): |
| """Retrieves the svn account which we assume is an email address.""" |
| infos = SVN.CaptureInfo(repo_root) |
| uuid = infos.get('UUID') |
| root = infos.get('Repository Root') |
| if not root: |
| return None |
| |
| # Should check for uuid but it is incorrectly saved for https creds. |
| realm = root.rsplit('/', 1)[0] |
| if root.startswith('https') or not uuid: |
| regexp = re.compile(r'<%s:\d+>.*' % realm) |
| else: |
| regexp = re.compile(r'<%s:\d+> %s' % (realm, uuid)) |
| if regexp is None: |
| return None |
| if sys.platform.startswith('win'): |
| if not 'APPDATA' in os.environ: |
| return None |
| auth_dir = os.path.join(os.environ['APPDATA'], 'Subversion', 'auth', |
| 'svn.simple') |
| else: |
| if not 'HOME' in os.environ: |
| return None |
| auth_dir = os.path.join(os.environ['HOME'], '.subversion', 'auth', |
| 'svn.simple') |
| for credfile in os.listdir(auth_dir): |
| cred_info = SVN.ReadSimpleAuth(os.path.join(auth_dir, credfile)) |
| if regexp.match(cred_info.get('svn:realmstring')): |
| return cred_info.get('username') |
| |
| @staticmethod |
| def ReadSimpleAuth(filename): |
| f = open(filename, 'r') |
| values = {} |
| def ReadOneItem(item_type): |
| m = re.match(r'%s (\d+)' % item_type, f.readline()) |
| if not m: |
| return None |
| data = f.read(int(m.group(1))) |
| if f.read(1) != '\n': |
| return None |
| return data |
| |
| while True: |
| key = ReadOneItem('K') |
| if not key: |
| break |
| value = ReadOneItem('V') |
| if not value: |
| break |
| values[key] = value |
| return values |
| |
| @staticmethod |
| def GetCheckoutRoot(directory): |
| """Returns the top level directory of the current repository. |
| |
| The directory is returned as an absolute path. |
| """ |
| directory = os.path.abspath(directory) |
| infos = SVN.CaptureInfo(directory, print_error=False) |
| cur_dir_repo_root = infos.get("Repository Root") |
| if not cur_dir_repo_root: |
| return None |
| |
| while True: |
| parent = os.path.dirname(directory) |
| if (SVN.CaptureInfo(parent, print_error=False).get( |
| "Repository Root") != cur_dir_repo_root): |
| break |
| directory = parent |
| return GetCasedPath(directory) |
| |
| @staticmethod |
| def AssertVersion(min_version): |
| """Asserts svn's version is at least min_version.""" |
| def only_int(val): |
| if val.isdigit(): |
| return int(val) |
| else: |
| return 0 |
| if not SVN.current_version: |
| SVN.current_version = SVN.Capture(['--version']).split()[2] |
| current_version_list = map(only_int, SVN.current_version.split('.')) |
| for min_ver in map(int, min_version.split('.')): |
| ver = current_version_list.pop(0) |
| if ver < min_ver: |
| return (False, SVN.current_version) |
| elif ver > min_ver: |
| return (True, SVN.current_version) |
| return (True, SVN.current_version) |