branch.c: use 'ref-filter' APIs
[gitweb.git] / shallow.c
index 9488edcbc00278525c6d75e38c1b39acffdceae5..4ded4a8cde45e01bdd2d551ccf45226e82f1ff4c 100644 (file)
--- a/shallow.c
+++ b/shallow.c
@@ -48,7 +48,7 @@ int is_repository_shallow(void)
                return is_shallow;
 
        if (!path)
-               path = git_path("shallow");
+               path = git_path_shallow();
        /*
         * fetch-pack sets '--shallow-file ""' as an indicator that no
         * shallow file should be used. We could just open it and it
@@ -142,7 +142,7 @@ static void check_shallow_file_for_update(void)
        if (is_shallow == -1)
                die("BUG: shallow must be initialized by now");
 
-       if (!stat_validity_check(&shallow_stat, git_path("shallow")))
+       if (!stat_validity_check(&shallow_stat, git_path_shallow()))
                die("shallow file has changed since we read it");
 }
 
@@ -261,7 +261,7 @@ void setup_alternate_shallow(struct lock_file *shallow_lock,
        struct strbuf sb = STRBUF_INIT;
        int fd;
 
-       fd = hold_lock_file_for_update(shallow_lock, git_path("shallow"),
+       fd = hold_lock_file_for_update(shallow_lock, git_path_shallow(),
                                       LOCK_DIE_ON_ERROR);
        check_shallow_file_for_update();
        if (write_shallow_commits(&sb, 0, extra)) {
@@ -308,7 +308,7 @@ void prune_shallow(int show_only)
                strbuf_release(&sb);
                return;
        }
-       fd = hold_lock_file_for_update(&shallow_lock, git_path("shallow"),
+       fd = hold_lock_file_for_update(&shallow_lock, git_path_shallow(),
                                       LOCK_DIE_ON_ERROR);
        check_shallow_file_for_update();
        if (write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY)) {
@@ -317,7 +317,7 @@ void prune_shallow(int show_only)
                                  shallow_lock.filename.buf);
                commit_lock_file(&shallow_lock);
        } else {
-               unlink(git_path("shallow"));
+               unlink(git_path_shallow());
                rollback_lock_file(&shallow_lock);
        }
        strbuf_release(&sb);
@@ -475,11 +475,10 @@ static void paint_down(struct paint_info *info, const unsigned char *sha1,
        free(tmp);
 }
 
-static int mark_uninteresting(const char *refname,
-                             const unsigned char *sha1,
+static int mark_uninteresting(const char *refname, const struct object_id *oid,
                              int flags, void *cb_data)
 {
-       struct commit *commit = lookup_commit_reference_gently(sha1, 1);
+       struct commit *commit = lookup_commit_reference_gently(oid->hash, 1);
        if (!commit)
                return 0;
        commit->object.flags |= UNINTERESTING;
@@ -512,8 +511,6 @@ void assign_shallow_commits_to_refs(struct shallow_info *info,
        unsigned int i, nr;
        int *shallow, nr_shallow = 0;
        struct paint_info pi;
-       struct each_ref_fn_sha1_adapter wrapped_mark_uninteresting =
-               {mark_uninteresting, NULL};
 
        trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
        shallow = xmalloc(sizeof(*shallow) * (info->nr_ours + info->nr_theirs));
@@ -544,8 +541,8 @@ void assign_shallow_commits_to_refs(struct shallow_info *info,
         * connect to old refs. If not (e.g. force ref updates) it'll
         * have to go down to the current shallow commits.
         */
-       head_ref(each_ref_fn_adapter, &wrapped_mark_uninteresting);
-       for_each_ref(each_ref_fn_adapter, &wrapped_mark_uninteresting);
+       head_ref(mark_uninteresting, NULL);
+       for_each_ref(mark_uninteresting, NULL);
 
        /* Mark potential bottoms so we won't go out of bound */
        for (i = 0; i < nr_shallow; i++) {
@@ -586,12 +583,12 @@ struct commit_array {
        int nr, alloc;
 };
 
-static int add_ref(const char *refname,
-                  const unsigned char *sha1, int flags, void *cb_data)
+static int add_ref(const char *refname, const struct object_id *oid,
+                  int flags, void *cb_data)
 {
        struct commit_array *ca = cb_data;
        ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
-       ca->commits[ca->nr] = lookup_commit_reference_gently(sha1, 1);
+       ca->commits[ca->nr] = lookup_commit_reference_gently(oid->hash, 1);
        if (ca->commits[ca->nr])
                ca->nr++;
        return 0;
@@ -620,8 +617,6 @@ static void post_assign_shallow(struct shallow_info *info,
        int dst, i, j;
        int bitmap_nr = (info->ref->nr + 31) / 32;
        struct commit_array ca;
-       struct each_ref_fn_sha1_adapter wrapped_add_ref =
-               {add_ref, &ca};
 
        trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
        if (ref_status)
@@ -645,8 +640,8 @@ static void post_assign_shallow(struct shallow_info *info,
        info->nr_theirs = dst;
 
        memset(&ca, 0, sizeof(ca));
-       head_ref(each_ref_fn_adapter, &wrapped_add_ref);
-       for_each_ref(each_ref_fn_adapter, &wrapped_add_ref);
+       head_ref(add_ref, &ca);
+       for_each_ref(add_ref, &ca);
 
        /* Remove unreachable shallow commits from "ours" */
        for (i = dst = 0; i < info->nr_ours; i++) {
@@ -678,12 +673,10 @@ int delayed_reachability_test(struct shallow_info *si, int c)
 
                if (!si->commits) {
                        struct commit_array ca;
-                       struct each_ref_fn_sha1_adapter wrapped_add_ref =
-                               {add_ref, &ca};
 
                        memset(&ca, 0, sizeof(ca));
-                       head_ref(each_ref_fn_adapter, &wrapped_add_ref);
-                       for_each_ref(each_ref_fn_adapter, &wrapped_add_ref);
+                       head_ref(add_ref, &ca);
+                       for_each_ref(add_ref, &ca);
                        si->commits = ca.commits;
                        si->nr_commits = ca.nr;
                }