commit: add repository argument to parse_commit_buffer
[git.git] / shallow.c
blobe9ce55bba08489d30a7a4516fc39e332260ac2f7
1 #include "cache.h"
2 #include "repository.h"
3 #include "tempfile.h"
4 #include "lockfile.h"
5 #include "object-store.h"
6 #include "commit.h"
7 #include "tag.h"
8 #include "pkt-line.h"
9 #include "remote.h"
10 #include "refs.h"
11 #include "sha1-array.h"
12 #include "diff.h"
13 #include "revision.h"
14 #include "commit-slab.h"
15 #include "revision.h"
16 #include "list-objects.h"
17 #include "commit-slab.h"
18 #include "repository.h"
20 void set_alternate_shallow_file(struct repository *r, const char *path, int override)
22 if (r->parsed_objects->is_shallow != -1)
23 BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
24 if (r->parsed_objects->alternate_shallow_file && !override)
25 return;
26 free(r->parsed_objects->alternate_shallow_file);
27 r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
30 int register_shallow(struct repository *r, const struct object_id *oid)
32 struct commit_graft *graft =
33 xmalloc(sizeof(struct commit_graft));
34 struct commit *commit = lookup_commit(the_repository, oid);
36 oidcpy(&graft->oid, oid);
37 graft->nr_parent = -1;
38 if (commit && commit->object.parsed)
39 commit->parents = NULL;
40 return register_commit_graft(r, graft, 0);
43 int is_repository_shallow(struct repository *r)
45 FILE *fp;
46 char buf[1024];
47 const char *path = r->parsed_objects->alternate_shallow_file;
49 if (r->parsed_objects->is_shallow >= 0)
50 return r->parsed_objects->is_shallow;
52 if (!path)
53 path = git_path_shallow(r);
55 * fetch-pack sets '--shallow-file ""' as an indicator that no
56 * shallow file should be used. We could just open it and it
57 * will likely fail. But let's do an explicit check instead.
59 if (!*path || (fp = fopen(path, "r")) == NULL) {
60 stat_validity_clear(r->parsed_objects->shallow_stat);
61 r->parsed_objects->is_shallow = 0;
62 return r->parsed_objects->is_shallow;
64 stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
65 r->parsed_objects->is_shallow = 1;
67 while (fgets(buf, sizeof(buf), fp)) {
68 struct object_id oid;
69 if (get_oid_hex(buf, &oid))
70 die("bad shallow line: %s", buf);
71 register_shallow(r, &oid);
73 fclose(fp);
74 return r->parsed_objects->is_shallow;
78 * TODO: use "int" elemtype instead of "int *" when/if commit-slab
79 * supports a "valid" flag.
81 define_commit_slab(commit_depth, int *);
82 struct commit_list *get_shallow_commits(struct object_array *heads, int depth,
83 int shallow_flag, int not_shallow_flag)
85 int i = 0, cur_depth = 0;
86 struct commit_list *result = NULL;
87 struct object_array stack = OBJECT_ARRAY_INIT;
88 struct commit *commit = NULL;
89 struct commit_graft *graft;
90 struct commit_depth depths;
92 init_commit_depth(&depths);
93 while (commit || i < heads->nr || stack.nr) {
94 struct commit_list *p;
95 if (!commit) {
96 if (i < heads->nr) {
97 int **depth_slot;
98 commit = (struct commit *)
99 deref_tag(heads->objects[i++].item, NULL, 0);
100 if (!commit || commit->object.type != OBJ_COMMIT) {
101 commit = NULL;
102 continue;
104 depth_slot = commit_depth_at(&depths, commit);
105 if (!*depth_slot)
106 *depth_slot = xmalloc(sizeof(int));
107 **depth_slot = 0;
108 cur_depth = 0;
109 } else {
110 commit = (struct commit *)
111 object_array_pop(&stack);
112 cur_depth = **commit_depth_at(&depths, commit);
115 parse_commit_or_die(commit);
116 cur_depth++;
117 if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
118 (is_repository_shallow(the_repository) && !commit->parents &&
119 (graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
120 graft->nr_parent < 0)) {
121 commit_list_insert(commit, &result);
122 commit->object.flags |= shallow_flag;
123 commit = NULL;
124 continue;
126 commit->object.flags |= not_shallow_flag;
127 for (p = commit->parents, commit = NULL; p; p = p->next) {
128 int **depth_slot = commit_depth_at(&depths, p->item);
129 if (!*depth_slot) {
130 *depth_slot = xmalloc(sizeof(int));
131 **depth_slot = cur_depth;
132 } else {
133 if (cur_depth >= **depth_slot)
134 continue;
135 **depth_slot = cur_depth;
137 if (p->next)
138 add_object_array(&p->item->object,
139 NULL, &stack);
140 else {
141 commit = p->item;
142 cur_depth = **commit_depth_at(&depths, commit);
146 for (i = 0; i < depths.slab_count; i++) {
147 int j;
149 for (j = 0; j < depths.slab_size; j++)
150 free(depths.slab[i][j]);
152 clear_commit_depth(&depths);
154 return result;
157 static void show_commit(struct commit *commit, void *data)
159 commit_list_insert(commit, data);
163 * Given rev-list arguments, run rev-list. All reachable commits
164 * except border ones are marked with not_shallow_flag. Border commits
165 * are marked with shallow_flag. The list of border/shallow commits
166 * are also returned.
168 struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av,
169 int shallow_flag,
170 int not_shallow_flag)
172 struct commit_list *result = NULL, *p;
173 struct commit_list *not_shallow_list = NULL;
174 struct rev_info revs;
175 int both_flags = shallow_flag | not_shallow_flag;
178 * SHALLOW (excluded) and NOT_SHALLOW (included) should not be
179 * set at this point. But better be safe than sorry.
181 clear_object_flags(both_flags);
183 is_repository_shallow(the_repository); /* make sure shallows are read */
185 init_revisions(&revs, NULL);
186 save_commit_buffer = 0;
187 setup_revisions(ac, av, &revs, NULL);
189 if (prepare_revision_walk(&revs))
190 die("revision walk setup failed");
191 traverse_commit_list(&revs, show_commit, NULL, &not_shallow_list);
193 if (!not_shallow_list)
194 die("no commits selected for shallow requests");
196 /* Mark all reachable commits as NOT_SHALLOW */
197 for (p = not_shallow_list; p; p = p->next)
198 p->item->object.flags |= not_shallow_flag;
201 * mark border commits SHALLOW + NOT_SHALLOW.
202 * We cannot clear NOT_SHALLOW right now. Imagine border
203 * commit A is processed first, then commit B, whose parent is
204 * A, later. If NOT_SHALLOW on A is cleared at step 1, B
205 * itself is considered border at step 2, which is incorrect.
207 for (p = not_shallow_list; p; p = p->next) {
208 struct commit *c = p->item;
209 struct commit_list *parent;
211 if (parse_commit(c))
212 die("unable to parse commit %s",
213 oid_to_hex(&c->object.oid));
215 for (parent = c->parents; parent; parent = parent->next)
216 if (!(parent->item->object.flags & not_shallow_flag)) {
217 c->object.flags |= shallow_flag;
218 commit_list_insert(c, &result);
219 break;
222 free_commit_list(not_shallow_list);
225 * Now we can clean up NOT_SHALLOW on border commits. Having
226 * both flags set can confuse the caller.
228 for (p = result; p; p = p->next) {
229 struct object *o = &p->item->object;
230 if ((o->flags & both_flags) == both_flags)
231 o->flags &= ~not_shallow_flag;
233 return result;
236 static void check_shallow_file_for_update(struct repository *r)
238 if (r->parsed_objects->is_shallow == -1)
239 BUG("shallow must be initialized by now");
241 if (!stat_validity_check(r->parsed_objects->shallow_stat, git_path_shallow(the_repository)))
242 die("shallow file has changed since we read it");
245 #define SEEN_ONLY 1
246 #define VERBOSE 2
248 struct write_shallow_data {
249 struct strbuf *out;
250 int use_pack_protocol;
251 int count;
252 unsigned flags;
255 static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
257 struct write_shallow_data *data = cb_data;
258 const char *hex = oid_to_hex(&graft->oid);
259 if (graft->nr_parent != -1)
260 return 0;
261 if (data->flags & SEEN_ONLY) {
262 struct commit *c = lookup_commit(the_repository, &graft->oid);
263 if (!c || !(c->object.flags & SEEN)) {
264 if (data->flags & VERBOSE)
265 printf("Removing %s from .git/shallow\n",
266 oid_to_hex(&c->object.oid));
267 return 0;
270 data->count++;
271 if (data->use_pack_protocol)
272 packet_buf_write(data->out, "shallow %s", hex);
273 else {
274 strbuf_addstr(data->out, hex);
275 strbuf_addch(data->out, '\n');
277 return 0;
280 static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
281 const struct oid_array *extra,
282 unsigned flags)
284 struct write_shallow_data data;
285 int i;
286 data.out = out;
287 data.use_pack_protocol = use_pack_protocol;
288 data.count = 0;
289 data.flags = flags;
290 for_each_commit_graft(write_one_shallow, &data);
291 if (!extra)
292 return data.count;
293 for (i = 0; i < extra->nr; i++) {
294 strbuf_addstr(out, oid_to_hex(extra->oid + i));
295 strbuf_addch(out, '\n');
296 data.count++;
298 return data.count;
301 int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
302 const struct oid_array *extra)
304 return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
307 const char *setup_temporary_shallow(const struct oid_array *extra)
309 struct tempfile *temp;
310 struct strbuf sb = STRBUF_INIT;
312 if (write_shallow_commits(&sb, 0, extra)) {
313 temp = xmks_tempfile(git_path("shallow_XXXXXX"));
315 if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
316 close_tempfile_gently(temp) < 0)
317 die_errno("failed to write to %s",
318 get_tempfile_path(temp));
319 strbuf_release(&sb);
320 return get_tempfile_path(temp);
323 * is_repository_shallow() sees empty string as "no shallow
324 * file".
326 return "";
329 void setup_alternate_shallow(struct lock_file *shallow_lock,
330 const char **alternate_shallow_file,
331 const struct oid_array *extra)
333 struct strbuf sb = STRBUF_INIT;
334 int fd;
336 fd = hold_lock_file_for_update(shallow_lock,
337 git_path_shallow(the_repository),
338 LOCK_DIE_ON_ERROR);
339 check_shallow_file_for_update(the_repository);
340 if (write_shallow_commits(&sb, 0, extra)) {
341 if (write_in_full(fd, sb.buf, sb.len) < 0)
342 die_errno("failed to write to %s",
343 get_lock_file_path(shallow_lock));
344 *alternate_shallow_file = get_lock_file_path(shallow_lock);
345 } else
347 * is_repository_shallow() sees empty string as "no
348 * shallow file".
350 *alternate_shallow_file = "";
351 strbuf_release(&sb);
354 static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
356 int fd = *(int *)cb;
357 if (graft->nr_parent == -1)
358 packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
359 return 0;
362 void advertise_shallow_grafts(int fd)
364 if (!is_repository_shallow(the_repository))
365 return;
366 for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
370 * mark_reachable_objects() should have been run prior to this and all
371 * reachable commits marked as "SEEN".
373 void prune_shallow(int show_only)
375 struct lock_file shallow_lock = LOCK_INIT;
376 struct strbuf sb = STRBUF_INIT;
377 int fd;
379 if (show_only) {
380 write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY | VERBOSE);
381 strbuf_release(&sb);
382 return;
384 fd = hold_lock_file_for_update(&shallow_lock,
385 git_path_shallow(the_repository),
386 LOCK_DIE_ON_ERROR);
387 check_shallow_file_for_update(the_repository);
388 if (write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY)) {
389 if (write_in_full(fd, sb.buf, sb.len) < 0)
390 die_errno("failed to write to %s",
391 get_lock_file_path(&shallow_lock));
392 commit_lock_file(&shallow_lock);
393 } else {
394 unlink(git_path_shallow(the_repository));
395 rollback_lock_file(&shallow_lock);
397 strbuf_release(&sb);
400 struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
403 * Step 1, split sender shallow commits into "ours" and "theirs"
404 * Step 2, clean "ours" based on .git/shallow
406 void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
408 int i;
409 trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
410 memset(info, 0, sizeof(*info));
411 info->shallow = sa;
412 if (!sa)
413 return;
414 ALLOC_ARRAY(info->ours, sa->nr);
415 ALLOC_ARRAY(info->theirs, sa->nr);
416 for (i = 0; i < sa->nr; i++) {
417 if (has_object_file(sa->oid + i)) {
418 struct commit_graft *graft;
419 graft = lookup_commit_graft(the_repository,
420 &sa->oid[i]);
421 if (graft && graft->nr_parent < 0)
422 continue;
423 info->ours[info->nr_ours++] = i;
424 } else
425 info->theirs[info->nr_theirs++] = i;
429 void clear_shallow_info(struct shallow_info *info)
431 free(info->ours);
432 free(info->theirs);
435 /* Step 4, remove non-existent ones in "theirs" after getting the pack */
437 void remove_nonexistent_theirs_shallow(struct shallow_info *info)
439 struct object_id *oid = info->shallow->oid;
440 int i, dst;
441 trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
442 for (i = dst = 0; i < info->nr_theirs; i++) {
443 if (i != dst)
444 info->theirs[dst] = info->theirs[i];
445 if (has_object_file(oid + info->theirs[i]))
446 dst++;
448 info->nr_theirs = dst;
451 define_commit_slab(ref_bitmap, uint32_t *);
453 #define POOL_SIZE (512 * 1024)
455 struct paint_info {
456 struct ref_bitmap ref_bitmap;
457 unsigned nr_bits;
458 char **pools;
459 char *free, *end;
460 unsigned pool_count;
463 static uint32_t *paint_alloc(struct paint_info *info)
465 unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
466 unsigned size = nr * sizeof(uint32_t);
467 void *p;
468 if (!info->pool_count || size > info->end - info->free) {
469 if (size > POOL_SIZE)
470 BUG("pool size too small for %d in paint_alloc()",
471 size);
472 info->pool_count++;
473 REALLOC_ARRAY(info->pools, info->pool_count);
474 info->free = xmalloc(POOL_SIZE);
475 info->pools[info->pool_count - 1] = info->free;
476 info->end = info->free + POOL_SIZE;
478 p = info->free;
479 info->free += size;
480 return p;
484 * Given a commit SHA-1, walk down to parents until either SEEN,
485 * UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
486 * all walked commits.
488 static void paint_down(struct paint_info *info, const struct object_id *oid,
489 unsigned int id)
491 unsigned int i, nr;
492 struct commit_list *head = NULL;
493 int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
494 size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
495 struct commit *c = lookup_commit_reference_gently(the_repository, oid,
497 uint32_t *tmp; /* to be freed before return */
498 uint32_t *bitmap;
500 if (!c)
501 return;
503 tmp = xmalloc(bitmap_size);
504 bitmap = paint_alloc(info);
505 memset(bitmap, 0, bitmap_size);
506 bitmap[id / 32] |= (1U << (id % 32));
507 commit_list_insert(c, &head);
508 while (head) {
509 struct commit_list *p;
510 struct commit *c = pop_commit(&head);
511 uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
513 /* XXX check "UNINTERESTING" from pack bitmaps if available */
514 if (c->object.flags & (SEEN | UNINTERESTING))
515 continue;
516 else
517 c->object.flags |= SEEN;
519 if (*refs == NULL)
520 *refs = bitmap;
521 else {
522 memcpy(tmp, *refs, bitmap_size);
523 for (i = 0; i < bitmap_nr; i++)
524 tmp[i] |= bitmap[i];
525 if (memcmp(tmp, *refs, bitmap_size)) {
526 *refs = paint_alloc(info);
527 memcpy(*refs, tmp, bitmap_size);
531 if (c->object.flags & BOTTOM)
532 continue;
534 if (parse_commit(c))
535 die("unable to parse commit %s",
536 oid_to_hex(&c->object.oid));
538 for (p = c->parents; p; p = p->next) {
539 if (p->item->object.flags & SEEN)
540 continue;
541 commit_list_insert(p->item, &head);
545 nr = get_max_object_index();
546 for (i = 0; i < nr; i++) {
547 struct object *o = get_indexed_object(i);
548 if (o && o->type == OBJ_COMMIT)
549 o->flags &= ~SEEN;
552 free(tmp);
555 static int mark_uninteresting(const char *refname, const struct object_id *oid,
556 int flags, void *cb_data)
558 struct commit *commit = lookup_commit_reference_gently(the_repository,
559 oid, 1);
560 if (!commit)
561 return 0;
562 commit->object.flags |= UNINTERESTING;
563 mark_parents_uninteresting(commit);
564 return 0;
567 static void post_assign_shallow(struct shallow_info *info,
568 struct ref_bitmap *ref_bitmap,
569 int *ref_status);
571 * Step 6(+7), associate shallow commits with new refs
573 * info->ref must be initialized before calling this function.
575 * If used is not NULL, it's an array of info->shallow->nr
576 * bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
577 * m-th shallow commit from info->shallow.
579 * If used is NULL, "ours" and "theirs" are updated. And if ref_status
580 * is not NULL it's an array of ref->nr ints. ref_status[i] is true if
581 * the ref needs some shallow commits from either info->ours or
582 * info->theirs.
584 void assign_shallow_commits_to_refs(struct shallow_info *info,
585 uint32_t **used, int *ref_status)
587 struct object_id *oid = info->shallow->oid;
588 struct oid_array *ref = info->ref;
589 unsigned int i, nr;
590 int *shallow, nr_shallow = 0;
591 struct paint_info pi;
593 trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
594 ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
595 for (i = 0; i < info->nr_ours; i++)
596 shallow[nr_shallow++] = info->ours[i];
597 for (i = 0; i < info->nr_theirs; i++)
598 shallow[nr_shallow++] = info->theirs[i];
601 * Prepare the commit graph to track what refs can reach what
602 * (new) shallow commits.
604 nr = get_max_object_index();
605 for (i = 0; i < nr; i++) {
606 struct object *o = get_indexed_object(i);
607 if (!o || o->type != OBJ_COMMIT)
608 continue;
610 o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
613 memset(&pi, 0, sizeof(pi));
614 init_ref_bitmap(&pi.ref_bitmap);
615 pi.nr_bits = ref->nr;
618 * "--not --all" to cut short the traversal if new refs
619 * connect to old refs. If not (e.g. force ref updates) it'll
620 * have to go down to the current shallow commits.
622 head_ref(mark_uninteresting, NULL);
623 for_each_ref(mark_uninteresting, NULL);
625 /* Mark potential bottoms so we won't go out of bound */
626 for (i = 0; i < nr_shallow; i++) {
627 struct commit *c = lookup_commit(the_repository,
628 &oid[shallow[i]]);
629 c->object.flags |= BOTTOM;
632 for (i = 0; i < ref->nr; i++)
633 paint_down(&pi, ref->oid + i, i);
635 if (used) {
636 int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
637 memset(used, 0, sizeof(*used) * info->shallow->nr);
638 for (i = 0; i < nr_shallow; i++) {
639 const struct commit *c = lookup_commit(the_repository,
640 &oid[shallow[i]]);
641 uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
642 if (*map)
643 used[shallow[i]] = xmemdupz(*map, bitmap_size);
646 * unreachable shallow commits are not removed from
647 * "ours" and "theirs". The user is supposed to run
648 * step 7 on every ref separately and not trust "ours"
649 * and "theirs" any more.
651 } else
652 post_assign_shallow(info, &pi.ref_bitmap, ref_status);
654 clear_ref_bitmap(&pi.ref_bitmap);
655 for (i = 0; i < pi.pool_count; i++)
656 free(pi.pools[i]);
657 free(pi.pools);
658 free(shallow);
661 struct commit_array {
662 struct commit **commits;
663 int nr, alloc;
666 static int add_ref(const char *refname, const struct object_id *oid,
667 int flags, void *cb_data)
669 struct commit_array *ca = cb_data;
670 ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
671 ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository,
672 oid, 1);
673 if (ca->commits[ca->nr])
674 ca->nr++;
675 return 0;
678 static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap)
680 unsigned int i;
681 if (!ref_status)
682 return;
683 for (i = 0; i < nr; i++)
684 if (bitmap[i / 32] & (1U << (i % 32)))
685 ref_status[i]++;
689 * Step 7, reachability test on "ours" at commit level
691 static void post_assign_shallow(struct shallow_info *info,
692 struct ref_bitmap *ref_bitmap,
693 int *ref_status)
695 struct object_id *oid = info->shallow->oid;
696 struct commit *c;
697 uint32_t **bitmap;
698 int dst, i, j;
699 int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
700 struct commit_array ca;
702 trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
703 if (ref_status)
704 memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr);
706 /* Remove unreachable shallow commits from "theirs" */
707 for (i = dst = 0; i < info->nr_theirs; i++) {
708 if (i != dst)
709 info->theirs[dst] = info->theirs[i];
710 c = lookup_commit(the_repository, &oid[info->theirs[i]]);
711 bitmap = ref_bitmap_at(ref_bitmap, c);
712 if (!*bitmap)
713 continue;
714 for (j = 0; j < bitmap_nr; j++)
715 if (bitmap[0][j]) {
716 update_refstatus(ref_status, info->ref->nr, *bitmap);
717 dst++;
718 break;
721 info->nr_theirs = dst;
723 memset(&ca, 0, sizeof(ca));
724 head_ref(add_ref, &ca);
725 for_each_ref(add_ref, &ca);
727 /* Remove unreachable shallow commits from "ours" */
728 for (i = dst = 0; i < info->nr_ours; i++) {
729 if (i != dst)
730 info->ours[dst] = info->ours[i];
731 c = lookup_commit(the_repository, &oid[info->ours[i]]);
732 bitmap = ref_bitmap_at(ref_bitmap, c);
733 if (!*bitmap)
734 continue;
735 for (j = 0; j < bitmap_nr; j++)
736 if (bitmap[0][j] &&
737 /* Step 7, reachability test at commit level */
738 !in_merge_bases_many(c, ca.nr, ca.commits)) {
739 update_refstatus(ref_status, info->ref->nr, *bitmap);
740 dst++;
741 break;
744 info->nr_ours = dst;
746 free(ca.commits);
749 /* (Delayed) step 7, reachability test at commit level */
750 int delayed_reachability_test(struct shallow_info *si, int c)
752 if (si->need_reachability_test[c]) {
753 struct commit *commit = lookup_commit(the_repository,
754 &si->shallow->oid[c]);
756 if (!si->commits) {
757 struct commit_array ca;
759 memset(&ca, 0, sizeof(ca));
760 head_ref(add_ref, &ca);
761 for_each_ref(add_ref, &ca);
762 si->commits = ca.commits;
763 si->nr_commits = ca.nr;
766 si->reachable[c] = in_merge_bases_many(commit,
767 si->nr_commits,
768 si->commits);
769 si->need_reachability_test[c] = 0;
771 return si->reachable[c];