+ object_array_clear(&ctx->revs->pending);
+}
+
+static void do_traverse(struct traversal_context *ctx)
+{
+ struct commit *commit;
+ struct strbuf csp; /* callee's scratch pad */
+ strbuf_init(&csp, PATH_MAX);
+
+ while ((commit = get_revision(ctx->revs)) != NULL) {
+ /*
+ * an uninteresting boundary commit may not have its tree
+ * parsed yet, but we are not going to show them anyway
+ */
+ if (!ctx->revs->tree_objects)
+ ; /* do not bother loading tree */
+ else if (get_commit_tree(commit)) {
+ struct tree *tree = get_commit_tree(commit);
+ tree->object.flags |= NOT_USER_GIVEN;
+ add_pending_tree(ctx->revs, tree);
+ } else if (commit->object.parsed) {
+ die(_("unable to load root tree for commit %s"),
+ oid_to_hex(&commit->object.oid));
+ }
+ ctx->show_commit(commit, ctx->show_data);
+
+ if (ctx->revs->tree_blobs_in_commit_order)
+ /*
+ * NEEDSWORK: Adding the tree and then flushing it here
+ * needs a reallocation for each commit. Can we pass the
+ * tree directory without allocation churn?
+ */
+ traverse_trees_and_blobs(ctx, &csp);
+ }
+ traverse_trees_and_blobs(ctx, &csp);
+ strbuf_release(&csp);
+}
+
+void traverse_commit_list(struct rev_info *revs,
+ show_commit_fn show_commit,
+ show_object_fn show_object,
+ void *show_data)
+{
+ struct traversal_context ctx;
+ ctx.revs = revs;
+ ctx.show_commit = show_commit;
+ ctx.show_object = show_object;
+ ctx.show_data = show_data;
+ ctx.filter = NULL;
+ do_traverse(&ctx);
+}
+
+void traverse_commit_list_filtered(
+ struct list_objects_filter_options *filter_options,
+ struct rev_info *revs,
+ show_commit_fn show_commit,
+ show_object_fn show_object,
+ void *show_data,
+ struct oidset *omitted)
+{
+ struct traversal_context ctx;
+
+ ctx.revs = revs;
+ ctx.show_object = show_object;
+ ctx.show_commit = show_commit;
+ ctx.show_data = show_data;
+ ctx.filter = list_objects_filter__init(omitted, filter_options);
+ do_traverse(&ctx);
+ list_objects_filter__free(ctx.filter);