From: Simon Glass <simon.glass@canonical.com> The three phases of decompose_mega_merge() all follow the same pattern: run git log, parse the output, filter out commits already in the database. Extract two helpers to reduce the duplication: - _mega_preadd(): pre-add the mega-merge commit as 'skipped' in the DB - _mega_get_batch(): common git-log/parse/filter pattern shared by all three phases This makes each phase a single call to _mega_get_batch() instead of repeated inline blocks. Co-developed-by: Claude Opus 4.6 <noreply@anthropic.com> Signed-off-by: Simon Glass <simon.glass@canonical.com> --- tools/pickman/control.py | 96 ++++++++++++++++++++++------------------ 1 file changed, 53 insertions(+), 43 deletions(-) diff --git a/tools/pickman/control.py b/tools/pickman/control.py index d3cb8a8fff4..19cd2813b64 100644 --- a/tools/pickman/control.py +++ b/tools/pickman/control.py @@ -964,6 +964,49 @@ def detect_sub_merges(merge_hash): return [line for line in out.split('\n') if line] +def _mega_preadd(dbs, merge_hash): + """Pre-add a mega-merge commit to the database as 'skipped'. + + This prevents the mega-merge from appearing as an orphan commit. + Does nothing if the commit already exists in the database. + """ + if dbs.commit_get(merge_hash): + return + + source_id = None + sources = dbs.source_get_all() + if sources: + source_id = dbs.source_get_id(sources[0][0]) + if source_id: + info = run_git(['log', '-1', '--format=%s|%an', merge_hash]) + parts = info.split('|', 1) + subject = parts[0] + author = parts[1] if len(parts) > 1 else '' + dbs.commit_add(merge_hash, source_id, subject, author, + status='skipped') + dbs.commit() + + +def _mega_get_batch(dbs, exclude_ref, include_ref): + """Fetch a batch of unprocessed commits between two refs. + + Runs git log for the range ^exclude_ref include_ref, parses the + output and filters out commits already in the database. + + Returns: + list: CommitInfo tuples for unprocessed commits, may be empty + """ + log_output = run_git([ + 'log', '--reverse', '--format=%H|%h|%an|%s|%P', + f'^{exclude_ref}', include_ref + ]) + if not log_output: + return [] + + all_commits = parse_log_output(log_output, has_parents=True) + return [c for c in all_commits if not dbs.commit_get(c.hash)] + + def decompose_mega_merge(dbs, prev_commit, merge_hash, sub_merges): """Return the next unprocessed batch from a mega-merge @@ -990,60 +1033,27 @@ def decompose_mega_merge(dbs, prev_commit, merge_hash, sub_merges): first_parent = parents[0] second_parent = parents[1] - # Pre-add the mega-merge commit itself as skipped - if not dbs.commit_get(merge_hash): - source_id = None - sources = dbs.source_get_all() - if sources: - source_id = dbs.source_get_id(sources[0][0]) - if source_id: - info = run_git(['log', '-1', '--format=%s|%an', merge_hash]) - parts = info.split('|', 1) - subject = parts[0] - author = parts[1] if len(parts) > 1 else '' - dbs.commit_add(merge_hash, source_id, subject, author, - status='skipped') - dbs.commit() + _mega_preadd(dbs, merge_hash) # Phase 1: mainline commits before the merge - log_output = run_git([ - 'log', '--reverse', '--format=%H|%h|%an|%s|%P', - f'{prev_commit}..{first_parent}' - ]) - if log_output: - all_commits = parse_log_output(log_output, has_parents=True) - commits = [c for c in all_commits if not dbs.commit_get(c.hash)] - if commits: - return commits, first_parent + commits = _mega_get_batch(dbs, prev_commit, first_parent) + if commits: + return commits, first_parent # Phase 2: sub-merge batches prev_sub = first_parent for sub_hash in sub_merges: - # Get commits for this sub-merge - log_output = run_git([ - 'log', '--reverse', '--format=%H|%h|%an|%s|%P', - f'^{prev_sub}', sub_hash - ]) - if log_output: - all_commits = parse_log_output(log_output, has_parents=True) - commits = [c for c in all_commits if not dbs.commit_get(c.hash)] - if commits: - return commits, None + commits = _mega_get_batch(dbs, prev_sub, sub_hash) + if commits: + return commits, None prev_sub = sub_hash # Phase 3: remainder after the last sub-merge last_sub = sub_merges[-1] if sub_merges else first_parent - log_output = run_git([ - 'log', '--reverse', '--format=%H|%h|%an|%s|%P', - f'^{last_sub}', second_parent - ]) - if log_output: - all_commits = parse_log_output(log_output, has_parents=True) - commits = [c for c in all_commits if not dbs.commit_get(c.hash)] - if commits: - return commits, None + commits = _mega_get_batch(dbs, last_sub, second_parent) + if commits: + return commits, None - # All done return [], None -- 2.43.0