def open_root(self, base_revnum, baton):
if self.old_inventory.root is None:
# First time the root is set
- file_id = generate_file_id(self.revid, "")
+ file_id = generate_file_id(self.source, self.revid, "")
self.dir_baserev[file_id] = []
else:
assert self.old_inventory.root.revision is not None
def _get_new_id(self, parent_id, new_path):
if self.id_map.has_key(new_path):
return self.id_map[new_path]
- return generate_file_id(self.revid, new_path)
+ return generate_file_id(self.source, self.revid, new_path)
def delete_entry(self, path, revnum, parent_id, pool):
path = path.decode("utf-8")
import sha
-from revids import (generate_svn_revision_id, parse_svn_revision_id,
- escape_svn_path)
+from revids import escape_svn_path
def generate_svn_file_id(uuid, revnum, branch, path):
"""Create a file id identifying a Subversion file.
return ret
-def generate_file_id(revid, path):
- (uuid, branch, revnum) = parse_svn_revision_id(revid)
- return generate_svn_file_id(uuid, revnum, branch, path)
+def generate_file_id(repos, revid, path):
+ (branch, revnum) = repos.lookup_revision_id(revid)
+ return generate_svn_file_id(repos.uuid, revnum, branch, path)
-def get_local_changes(paths, scheme, uuid, get_children=None):
+def get_local_changes(paths, scheme, generate_revid, get_children=None):
new_paths = {}
names = paths.keys()
names.sort()
if (crp == "" and new_p == ""):
data = ('M', None, None)
else:
- data = (data[0], crp, generate_svn_revision_id(
- uuid, data[2], cbp))
+ data = (data[0], crp, generate_revid(data[2], cbp))
except NotBranchError:
# Copied from outside of a known branch
# Make it look like the files were added in this revision
:param global_changes: Dict with global changes that happened
"""
changes = get_local_changes(global_changes, self.repos.scheme,
- uuid, find_children)
+ self.repos.generate_revision_id, find_children)
if find_children is not None:
def get_children(path, revid):
(bp, revnum) = self.repos.lookup_revision_id(revid)
else:
get_children = None
- revid = generate_svn_revision_id(uuid, revnum, branch)
+ revid = self.repos.generate_revision_id(revnum, branch)
def new_file_id(x):
if renames.has_key(x):
return renames[x]
- return generate_file_id(revid, x)
+ return generate_file_id(self.repos, revid, x)
return self._apply_changes(new_file_id, changes, get_children)
# No history -> empty map
for (bp, paths, rev) in self.repos.follow_branch_history(branch, revnum):
- revid = generate_svn_revision_id(uuid, rev, bp)
+ revid = self.repos.generate_revision_id(rev, bp)
map = self.load(revid)
if map != {}:
# found the nearest cached map
def new_file_id(x):
if renames.has_key(x):
return renames[x]
- return generate_file_id(revid, x)
+ return generate_file_id(self.repos, revid, x)
revmap = self._apply_changes(new_file_id, changes, find_children)
for p in changes:
repos = self.olddir.find_repository()
inv = repos.get_inventory(repos.generate_revision_id(2, ""))
- import pdb
- pdb.set_trace()
self.assertEqual(repos.generate_revision_id(2, ""),
inv[inv.path2id('foo/bla')].revision)
- self.assertTrue(wt.branch.last_revision() in
- repos.revision_parents(repos.generate_revision_id(2, "")))
+ self.assertEqual(wt.branch.last_revision(),
+ repos.generate_revision_id(2, ""))
self.assertEqual(wt.branch.last_revision(),
self.olddir.open_branch().last_revision())
self.assertEqual("other data",
inv = repos.get_inventory(repos.generate_revision_id(2, ""))
self.assertEqual(repos.generate_revision_id(2, ""),
inv[inv.path2id('foo/bla')].revision)
- self.assertTrue(wt.branch.last_revision() in
- repos.revision_parents(repos.generate_revision_id(2, "")))
+ self.assertEqual(wt.branch.last_revision(),
+ repos.generate_revision_id(2, ""))
self.assertEqual(repos.generate_revision_id(2, ""),
self.svndir.open_branch().last_revision())
self.assertEqual("other data",