141f02b9c8933c0908f4e35ea141c3b32b39ce18
1#include "../cache.h"
2#include "../config.h"
3#include "../refs.h"
4#include "refs-internal.h"
5#include "ref-cache.h"
6#include "packed-backend.h"
7#include "../iterator.h"
8#include "../lockfile.h"
9
10struct packed_ref_store;
11
12struct packed_ref_cache {
13 /*
14 * A back-pointer to the packed_ref_store with which this
15 * cache is associated:
16 */
17 struct packed_ref_store *refs;
18
19 struct ref_cache *cache;
20
21 /*
22 * Count of references to the data structure in this instance,
23 * including the pointer from files_ref_store::packed if any.
24 * The data will not be freed as long as the reference count
25 * is nonzero.
26 */
27 unsigned int referrers;
28
29 /* The metadata from when this packed-refs cache was read */
30 struct stat_validity validity;
31};
32
33/*
34 * Increment the reference count of *packed_refs.
35 */
36static void acquire_packed_ref_cache(struct packed_ref_cache *packed_refs)
37{
38 packed_refs->referrers++;
39}
40
41/*
42 * Decrease the reference count of *packed_refs. If it goes to zero,
43 * free *packed_refs and return true; otherwise return false.
44 */
45static int release_packed_ref_cache(struct packed_ref_cache *packed_refs)
46{
47 if (!--packed_refs->referrers) {
48 free_ref_cache(packed_refs->cache);
49 stat_validity_clear(&packed_refs->validity);
50 free(packed_refs);
51 return 1;
52 } else {
53 return 0;
54 }
55}
56
57/*
58 * A container for `packed-refs`-related data. It is not (yet) a
59 * `ref_store`.
60 */
61struct packed_ref_store {
62 struct ref_store base;
63
64 unsigned int store_flags;
65
66 /* The path of the "packed-refs" file: */
67 char *path;
68
69 /*
70 * A cache of the values read from the `packed-refs` file, if
71 * it might still be current; otherwise, NULL.
72 */
73 struct packed_ref_cache *cache;
74
75 /*
76 * Lock used for the "packed-refs" file. Note that this (and
77 * thus the enclosing `packed_ref_store`) must not be freed.
78 */
79 struct lock_file lock;
80
81 /*
82 * Temporary file used when rewriting new contents to the
83 * "packed-refs" file. Note that this (and thus the enclosing
84 * `packed_ref_store`) must not be freed.
85 */
86 struct tempfile tempfile;
87};
88
89struct ref_store *packed_ref_store_create(const char *path,
90 unsigned int store_flags)
91{
92 struct packed_ref_store *refs = xcalloc(1, sizeof(*refs));
93 struct ref_store *ref_store = (struct ref_store *)refs;
94
95 base_ref_store_init(ref_store, &refs_be_packed);
96 refs->store_flags = store_flags;
97
98 refs->path = xstrdup(path);
99 return ref_store;
100}
101
102/*
103 * Downcast `ref_store` to `packed_ref_store`. Die if `ref_store` is
104 * not a `packed_ref_store`. Also die if `packed_ref_store` doesn't
105 * support at least the flags specified in `required_flags`. `caller`
106 * is used in any necessary error messages.
107 */
108static struct packed_ref_store *packed_downcast(struct ref_store *ref_store,
109 unsigned int required_flags,
110 const char *caller)
111{
112 struct packed_ref_store *refs;
113
114 if (ref_store->be != &refs_be_packed)
115 die("BUG: ref_store is type \"%s\" not \"packed\" in %s",
116 ref_store->be->name, caller);
117
118 refs = (struct packed_ref_store *)ref_store;
119
120 if ((refs->store_flags & required_flags) != required_flags)
121 die("BUG: unallowed operation (%s), requires %x, has %x\n",
122 caller, required_flags, refs->store_flags);
123
124 return refs;
125}
126
127static void clear_packed_ref_cache(struct packed_ref_store *refs)
128{
129 if (refs->cache) {
130 struct packed_ref_cache *cache = refs->cache;
131
132 refs->cache = NULL;
133 release_packed_ref_cache(cache);
134 }
135}
136
137/* The length of a peeled reference line in packed-refs, including EOL: */
138#define PEELED_LINE_LENGTH 42
139
140/*
141 * Parse one line from a packed-refs file. Write the SHA1 to sha1.
142 * Return a pointer to the refname within the line (null-terminated),
143 * or NULL if there was a problem.
144 */
145static const char *parse_ref_line(struct strbuf *line, struct object_id *oid)
146{
147 const char *ref;
148
149 if (parse_oid_hex(line->buf, oid, &ref) < 0)
150 return NULL;
151 if (!isspace(*ref++))
152 return NULL;
153
154 if (isspace(*ref))
155 return NULL;
156
157 if (line->buf[line->len - 1] != '\n')
158 return NULL;
159 line->buf[--line->len] = 0;
160
161 return ref;
162}
163
164static NORETURN void die_unterminated_line(const char *path,
165 const char *p, size_t len)
166{
167 if (len < 80)
168 die("unterminated line in %s: %.*s", path, (int)len, p);
169 else
170 die("unterminated line in %s: %.75s...", path, p);
171}
172
173static NORETURN void die_invalid_line(const char *path,
174 const char *p, size_t len)
175{
176 const char *eol = memchr(p, '\n', len);
177
178 if (!eol)
179 die_unterminated_line(path, p, len);
180 else if (eol - p < 80)
181 die("unexpected line in %s: %.*s", path, (int)(eol - p), p);
182 else
183 die("unexpected line in %s: %.75s...", path, p);
184
185}
186
187/*
188 * Read from the `packed-refs` file into a newly-allocated
189 * `packed_ref_cache` and return it. The return value will already
190 * have its reference count incremented.
191 *
192 * A comment line of the form "# pack-refs with: " may contain zero or
193 * more traits. We interpret the traits as follows:
194 *
195 * No traits:
196 *
197 * Probably no references are peeled. But if the file contains a
198 * peeled value for a reference, we will use it.
199 *
200 * peeled:
201 *
202 * References under "refs/tags/", if they *can* be peeled, *are*
203 * peeled in this file. References outside of "refs/tags/" are
204 * probably not peeled even if they could have been, but if we find
205 * a peeled value for such a reference we will use it.
206 *
207 * fully-peeled:
208 *
209 * All references in the file that can be peeled are peeled.
210 * Inversely (and this is more important), any references in the
211 * file for which no peeled value is recorded is not peelable. This
212 * trait should typically be written alongside "peeled" for
213 * compatibility with older clients, but we do not require it
214 * (i.e., "peeled" is a no-op if "fully-peeled" is set).
215 */
216static struct packed_ref_cache *read_packed_refs(struct packed_ref_store *refs)
217{
218 struct packed_ref_cache *packed_refs = xcalloc(1, sizeof(*packed_refs));
219 int fd;
220 struct stat st;
221 size_t size;
222 char *buf;
223 const char *pos, *eol, *eof;
224 struct ref_entry *last = NULL;
225 struct strbuf line = STRBUF_INIT;
226 enum { PEELED_NONE, PEELED_TAGS, PEELED_FULLY } peeled = PEELED_NONE;
227 struct ref_dir *dir;
228
229 packed_refs->refs = refs;
230 acquire_packed_ref_cache(packed_refs);
231 packed_refs->cache = create_ref_cache(NULL, NULL);
232 packed_refs->cache->root->flag &= ~REF_INCOMPLETE;
233
234 fd = open(refs->path, O_RDONLY);
235 if (fd < 0) {
236 if (errno == ENOENT) {
237 /*
238 * This is OK; it just means that no
239 * "packed-refs" file has been written yet,
240 * which is equivalent to it being empty.
241 */
242 return packed_refs;
243 } else {
244 die_errno("couldn't read %s", refs->path);
245 }
246 }
247
248 stat_validity_update(&packed_refs->validity, fd);
249
250 if (fstat(fd, &st) < 0)
251 die_errno("couldn't stat %s", refs->path);
252
253 size = xsize_t(st.st_size);
254 buf = xmmap(NULL, size, PROT_READ, MAP_PRIVATE, fd, 0);
255 pos = buf;
256 eof = buf + size;
257
258 /* If the file has a header line, process it: */
259 if (pos < eof && *pos == '#') {
260 const char *traits;
261
262 eol = memchr(pos, '\n', eof - pos);
263 if (!eol)
264 die_unterminated_line(refs->path, pos, eof - pos);
265
266 strbuf_add(&line, pos, eol + 1 - pos);
267
268 if (!skip_prefix(line.buf, "# pack-refs with:", &traits))
269 die_invalid_line(refs->path, pos, eof - pos);
270
271 if (strstr(traits, " fully-peeled "))
272 peeled = PEELED_FULLY;
273 else if (strstr(traits, " peeled "))
274 peeled = PEELED_TAGS;
275 /* perhaps other traits later as well */
276
277 /* The "+ 1" is for the LF character. */
278 pos = eol + 1;
279 strbuf_reset(&line);
280 }
281
282 dir = get_ref_dir(packed_refs->cache->root);
283 while (pos < eof) {
284 struct object_id oid;
285 const char *refname;
286
287 eol = memchr(pos, '\n', eof - pos);
288 if (!eol)
289 die_unterminated_line(refs->path, pos, eof - pos);
290
291 strbuf_add(&line, pos, eol + 1 - pos);
292
293 refname = parse_ref_line(&line, &oid);
294 if (refname) {
295 int flag = REF_ISPACKED;
296
297 if (check_refname_format(refname, REFNAME_ALLOW_ONELEVEL)) {
298 if (!refname_is_safe(refname))
299 die("packed refname is dangerous: %s", refname);
300 oidclr(&oid);
301 flag |= REF_BAD_NAME | REF_ISBROKEN;
302 }
303 last = create_ref_entry(refname, &oid, flag);
304 if (peeled == PEELED_FULLY ||
305 (peeled == PEELED_TAGS && starts_with(refname, "refs/tags/")))
306 last->flag |= REF_KNOWS_PEELED;
307 add_ref_entry(dir, last);
308 } else if (last &&
309 line.buf[0] == '^' &&
310 line.len == PEELED_LINE_LENGTH &&
311 line.buf[PEELED_LINE_LENGTH - 1] == '\n' &&
312 !get_oid_hex(line.buf + 1, &oid)) {
313 oidcpy(&last->u.value.peeled, &oid);
314 /*
315 * Regardless of what the file header said,
316 * we definitely know the value of *this*
317 * reference:
318 */
319 last->flag |= REF_KNOWS_PEELED;
320 } else {
321 die_invalid_line(refs->path, line.buf, line.len);
322 }
323
324 /* The "+ 1" is for the LF character. */
325 pos = eol + 1;
326 strbuf_reset(&line);
327 }
328
329 if (munmap(buf, size))
330 die_errno("error ummapping packed-refs file");
331 close(fd);
332
333 strbuf_release(&line);
334 return packed_refs;
335}
336
337/*
338 * Check that the packed refs cache (if any) still reflects the
339 * contents of the file. If not, clear the cache.
340 */
341static void validate_packed_ref_cache(struct packed_ref_store *refs)
342{
343 if (refs->cache &&
344 !stat_validity_check(&refs->cache->validity, refs->path))
345 clear_packed_ref_cache(refs);
346}
347
348/*
349 * Get the packed_ref_cache for the specified packed_ref_store,
350 * creating and populating it if it hasn't been read before or if the
351 * file has been changed (according to its `validity` field) since it
352 * was last read. On the other hand, if we hold the lock, then assume
353 * that the file hasn't been changed out from under us, so skip the
354 * extra `stat()` call in `stat_validity_check()`.
355 */
356static struct packed_ref_cache *get_packed_ref_cache(struct packed_ref_store *refs)
357{
358 if (!is_lock_file_locked(&refs->lock))
359 validate_packed_ref_cache(refs);
360
361 if (!refs->cache)
362 refs->cache = read_packed_refs(refs);
363
364 return refs->cache;
365}
366
367static struct ref_dir *get_packed_ref_dir(struct packed_ref_cache *packed_ref_cache)
368{
369 return get_ref_dir(packed_ref_cache->cache->root);
370}
371
372static struct ref_dir *get_packed_refs(struct packed_ref_store *refs)
373{
374 return get_packed_ref_dir(get_packed_ref_cache(refs));
375}
376
377/*
378 * Return the ref_entry for the given refname from the packed
379 * references. If it does not exist, return NULL.
380 */
381static struct ref_entry *get_packed_ref(struct packed_ref_store *refs,
382 const char *refname)
383{
384 return find_ref_entry(get_packed_refs(refs), refname);
385}
386
387static int packed_read_raw_ref(struct ref_store *ref_store,
388 const char *refname, unsigned char *sha1,
389 struct strbuf *referent, unsigned int *type)
390{
391 struct packed_ref_store *refs =
392 packed_downcast(ref_store, REF_STORE_READ, "read_raw_ref");
393
394 struct ref_entry *entry;
395
396 *type = 0;
397
398 entry = get_packed_ref(refs, refname);
399 if (!entry) {
400 errno = ENOENT;
401 return -1;
402 }
403
404 hashcpy(sha1, entry->u.value.oid.hash);
405 *type = REF_ISPACKED;
406 return 0;
407}
408
409static int packed_peel_ref(struct ref_store *ref_store,
410 const char *refname, unsigned char *sha1)
411{
412 struct packed_ref_store *refs =
413 packed_downcast(ref_store, REF_STORE_READ | REF_STORE_ODB,
414 "peel_ref");
415 struct ref_entry *r = get_packed_ref(refs, refname);
416
417 if (!r || peel_entry(r, 0))
418 return -1;
419
420 hashcpy(sha1, r->u.value.peeled.hash);
421 return 0;
422}
423
424struct packed_ref_iterator {
425 struct ref_iterator base;
426
427 struct packed_ref_cache *cache;
428 struct ref_iterator *iter0;
429 unsigned int flags;
430};
431
432static int packed_ref_iterator_advance(struct ref_iterator *ref_iterator)
433{
434 struct packed_ref_iterator *iter =
435 (struct packed_ref_iterator *)ref_iterator;
436 int ok;
437
438 while ((ok = ref_iterator_advance(iter->iter0)) == ITER_OK) {
439 if (iter->flags & DO_FOR_EACH_PER_WORKTREE_ONLY &&
440 ref_type(iter->iter0->refname) != REF_TYPE_PER_WORKTREE)
441 continue;
442
443 if (!(iter->flags & DO_FOR_EACH_INCLUDE_BROKEN) &&
444 !ref_resolves_to_object(iter->iter0->refname,
445 iter->iter0->oid,
446 iter->iter0->flags))
447 continue;
448
449 iter->base.refname = iter->iter0->refname;
450 iter->base.oid = iter->iter0->oid;
451 iter->base.flags = iter->iter0->flags;
452 return ITER_OK;
453 }
454
455 iter->iter0 = NULL;
456 if (ref_iterator_abort(ref_iterator) != ITER_DONE)
457 ok = ITER_ERROR;
458
459 return ok;
460}
461
462static int packed_ref_iterator_peel(struct ref_iterator *ref_iterator,
463 struct object_id *peeled)
464{
465 struct packed_ref_iterator *iter =
466 (struct packed_ref_iterator *)ref_iterator;
467
468 return ref_iterator_peel(iter->iter0, peeled);
469}
470
471static int packed_ref_iterator_abort(struct ref_iterator *ref_iterator)
472{
473 struct packed_ref_iterator *iter =
474 (struct packed_ref_iterator *)ref_iterator;
475 int ok = ITER_DONE;
476
477 if (iter->iter0)
478 ok = ref_iterator_abort(iter->iter0);
479
480 release_packed_ref_cache(iter->cache);
481 base_ref_iterator_free(ref_iterator);
482 return ok;
483}
484
485static struct ref_iterator_vtable packed_ref_iterator_vtable = {
486 packed_ref_iterator_advance,
487 packed_ref_iterator_peel,
488 packed_ref_iterator_abort
489};
490
491static struct ref_iterator *packed_ref_iterator_begin(
492 struct ref_store *ref_store,
493 const char *prefix, unsigned int flags)
494{
495 struct packed_ref_store *refs;
496 struct packed_ref_iterator *iter;
497 struct ref_iterator *ref_iterator;
498 unsigned int required_flags = REF_STORE_READ;
499
500 if (!(flags & DO_FOR_EACH_INCLUDE_BROKEN))
501 required_flags |= REF_STORE_ODB;
502 refs = packed_downcast(ref_store, required_flags, "ref_iterator_begin");
503
504 iter = xcalloc(1, sizeof(*iter));
505 ref_iterator = &iter->base;
506 base_ref_iterator_init(ref_iterator, &packed_ref_iterator_vtable, 1);
507
508 /*
509 * Note that get_packed_ref_cache() internally checks whether
510 * the packed-ref cache is up to date with what is on disk,
511 * and re-reads it if not.
512 */
513
514 iter->cache = get_packed_ref_cache(refs);
515 acquire_packed_ref_cache(iter->cache);
516 iter->iter0 = cache_ref_iterator_begin(iter->cache->cache, prefix, 0);
517
518 iter->flags = flags;
519
520 return ref_iterator;
521}
522
523/*
524 * Write an entry to the packed-refs file for the specified refname.
525 * If peeled is non-NULL, write it as the entry's peeled value. On
526 * error, return a nonzero value and leave errno set at the value left
527 * by the failing call to `fprintf()`.
528 */
529static int write_packed_entry(FILE *fh, const char *refname,
530 const unsigned char *sha1,
531 const unsigned char *peeled)
532{
533 if (fprintf(fh, "%s %s\n", sha1_to_hex(sha1), refname) < 0 ||
534 (peeled && fprintf(fh, "^%s\n", sha1_to_hex(peeled)) < 0))
535 return -1;
536
537 return 0;
538}
539
540int packed_refs_lock(struct ref_store *ref_store, int flags, struct strbuf *err)
541{
542 struct packed_ref_store *refs =
543 packed_downcast(ref_store, REF_STORE_WRITE | REF_STORE_MAIN,
544 "packed_refs_lock");
545 static int timeout_configured = 0;
546 static int timeout_value = 1000;
547
548 if (!timeout_configured) {
549 git_config_get_int("core.packedrefstimeout", &timeout_value);
550 timeout_configured = 1;
551 }
552
553 /*
554 * Note that we close the lockfile immediately because we
555 * don't write new content to it, but rather to a separate
556 * tempfile.
557 */
558 if (hold_lock_file_for_update_timeout(
559 &refs->lock,
560 refs->path,
561 flags, timeout_value) < 0) {
562 unable_to_lock_message(refs->path, errno, err);
563 return -1;
564 }
565
566 if (close_lock_file(&refs->lock)) {
567 strbuf_addf(err, "unable to close %s: %s", refs->path, strerror(errno));
568 return -1;
569 }
570
571 /*
572 * Now that we hold the `packed-refs` lock, make sure that our
573 * cache matches the current version of the file. Normally
574 * `get_packed_ref_cache()` does that for us, but that
575 * function assumes that when the file is locked, any existing
576 * cache is still valid. We've just locked the file, but it
577 * might have changed the moment *before* we locked it.
578 */
579 validate_packed_ref_cache(refs);
580
581 /*
582 * Now make sure that the packed-refs file as it exists in the
583 * locked state is loaded into the cache:
584 */
585 get_packed_ref_cache(refs);
586 return 0;
587}
588
589void packed_refs_unlock(struct ref_store *ref_store)
590{
591 struct packed_ref_store *refs = packed_downcast(
592 ref_store,
593 REF_STORE_READ | REF_STORE_WRITE,
594 "packed_refs_unlock");
595
596 if (!is_lock_file_locked(&refs->lock))
597 die("BUG: packed_refs_unlock() called when not locked");
598 rollback_lock_file(&refs->lock);
599}
600
601int packed_refs_is_locked(struct ref_store *ref_store)
602{
603 struct packed_ref_store *refs = packed_downcast(
604 ref_store,
605 REF_STORE_READ | REF_STORE_WRITE,
606 "packed_refs_is_locked");
607
608 return is_lock_file_locked(&refs->lock);
609}
610
611/*
612 * The packed-refs header line that we write out. Perhaps other
613 * traits will be added later. The trailing space is required.
614 */
615static const char PACKED_REFS_HEADER[] =
616 "# pack-refs with: peeled fully-peeled \n";
617
618static int packed_init_db(struct ref_store *ref_store, struct strbuf *err)
619{
620 /* Nothing to do. */
621 return 0;
622}
623
624/*
625 * Write the packed-refs from the cache to the packed-refs tempfile,
626 * incorporating any changes from `updates`. `updates` must be a
627 * sorted string list whose keys are the refnames and whose util
628 * values are `struct ref_update *`. On error, rollback the tempfile,
629 * write an error message to `err`, and return a nonzero value.
630 *
631 * The packfile must be locked before calling this function and will
632 * remain locked when it is done.
633 */
634static int write_with_updates(struct packed_ref_store *refs,
635 struct string_list *updates,
636 struct strbuf *err)
637{
638 struct ref_iterator *iter = NULL;
639 size_t i;
640 int ok;
641 FILE *out;
642 struct strbuf sb = STRBUF_INIT;
643 char *packed_refs_path;
644
645 if (!is_lock_file_locked(&refs->lock))
646 die("BUG: write_with_updates() called while unlocked");
647
648 /*
649 * If packed-refs is a symlink, we want to overwrite the
650 * symlinked-to file, not the symlink itself. Also, put the
651 * staging file next to it:
652 */
653 packed_refs_path = get_locked_file_path(&refs->lock);
654 strbuf_addf(&sb, "%s.new", packed_refs_path);
655 free(packed_refs_path);
656 if (create_tempfile(&refs->tempfile, sb.buf) < 0) {
657 strbuf_addf(err, "unable to create file %s: %s",
658 sb.buf, strerror(errno));
659 strbuf_release(&sb);
660 return -1;
661 }
662 strbuf_release(&sb);
663
664 out = fdopen_tempfile(&refs->tempfile, "w");
665 if (!out) {
666 strbuf_addf(err, "unable to fdopen packed-refs tempfile: %s",
667 strerror(errno));
668 goto error;
669 }
670
671 if (fprintf(out, "%s", PACKED_REFS_HEADER) < 0)
672 goto write_error;
673
674 /*
675 * We iterate in parallel through the current list of refs and
676 * the list of updates, processing an entry from at least one
677 * of the lists each time through the loop. When the current
678 * list of refs is exhausted, set iter to NULL. When the list
679 * of updates is exhausted, leave i set to updates->nr.
680 */
681 iter = packed_ref_iterator_begin(&refs->base, "",
682 DO_FOR_EACH_INCLUDE_BROKEN);
683 if ((ok = ref_iterator_advance(iter)) != ITER_OK)
684 iter = NULL;
685
686 i = 0;
687
688 while (iter || i < updates->nr) {
689 struct ref_update *update = NULL;
690 int cmp;
691
692 if (i >= updates->nr) {
693 cmp = -1;
694 } else {
695 update = updates->items[i].util;
696
697 if (!iter)
698 cmp = +1;
699 else
700 cmp = strcmp(iter->refname, update->refname);
701 }
702
703 if (!cmp) {
704 /*
705 * There is both an old value and an update
706 * for this reference. Check the old value if
707 * necessary:
708 */
709 if ((update->flags & REF_HAVE_OLD)) {
710 if (is_null_oid(&update->old_oid)) {
711 strbuf_addf(err, "cannot update ref '%s': "
712 "reference already exists",
713 update->refname);
714 goto error;
715 } else if (oidcmp(&update->old_oid, iter->oid)) {
716 strbuf_addf(err, "cannot update ref '%s': "
717 "is at %s but expected %s",
718 update->refname,
719 oid_to_hex(iter->oid),
720 oid_to_hex(&update->old_oid));
721 goto error;
722 }
723 }
724
725 /* Now figure out what to use for the new value: */
726 if ((update->flags & REF_HAVE_NEW)) {
727 /*
728 * The update takes precedence. Skip
729 * the iterator over the unneeded
730 * value.
731 */
732 if ((ok = ref_iterator_advance(iter)) != ITER_OK)
733 iter = NULL;
734 cmp = +1;
735 } else {
736 /*
737 * The update doesn't actually want to
738 * change anything. We're done with it.
739 */
740 i++;
741 cmp = -1;
742 }
743 } else if (cmp > 0) {
744 /*
745 * There is no old value but there is an
746 * update for this reference. Make sure that
747 * the update didn't expect an existing value:
748 */
749 if ((update->flags & REF_HAVE_OLD) &&
750 !is_null_oid(&update->old_oid)) {
751 strbuf_addf(err, "cannot update ref '%s': "
752 "reference is missing but expected %s",
753 update->refname,
754 oid_to_hex(&update->old_oid));
755 goto error;
756 }
757 }
758
759 if (cmp < 0) {
760 /* Pass the old reference through. */
761
762 struct object_id peeled;
763 int peel_error = ref_iterator_peel(iter, &peeled);
764
765 if (write_packed_entry(out, iter->refname,
766 iter->oid->hash,
767 peel_error ? NULL : peeled.hash))
768 goto write_error;
769
770 if ((ok = ref_iterator_advance(iter)) != ITER_OK)
771 iter = NULL;
772 } else if (is_null_oid(&update->new_oid)) {
773 /*
774 * The update wants to delete the reference,
775 * and the reference either didn't exist or we
776 * have already skipped it. So we're done with
777 * the update (and don't have to write
778 * anything).
779 */
780 i++;
781 } else {
782 struct object_id peeled;
783 int peel_error = peel_object(update->new_oid.hash,
784 peeled.hash);
785
786 if (write_packed_entry(out, update->refname,
787 update->new_oid.hash,
788 peel_error ? NULL : peeled.hash))
789 goto write_error;
790
791 i++;
792 }
793 }
794
795 if (ok != ITER_DONE) {
796 strbuf_addf(err, "unable to write packed-refs file: "
797 "error iterating over old contents");
798 goto error;
799 }
800
801 if (close_tempfile(&refs->tempfile)) {
802 strbuf_addf(err, "error closing file %s: %s",
803 get_tempfile_path(&refs->tempfile),
804 strerror(errno));
805 strbuf_release(&sb);
806 return -1;
807 }
808
809 return 0;
810
811write_error:
812 strbuf_addf(err, "error writing to %s: %s",
813 get_tempfile_path(&refs->tempfile), strerror(errno));
814
815error:
816 if (iter)
817 ref_iterator_abort(iter);
818
819 delete_tempfile(&refs->tempfile);
820 return -1;
821}
822
823struct packed_transaction_backend_data {
824 /* True iff the transaction owns the packed-refs lock. */
825 int own_lock;
826
827 struct string_list updates;
828};
829
830static void packed_transaction_cleanup(struct packed_ref_store *refs,
831 struct ref_transaction *transaction)
832{
833 struct packed_transaction_backend_data *data = transaction->backend_data;
834
835 if (data) {
836 string_list_clear(&data->updates, 0);
837
838 if (is_tempfile_active(&refs->tempfile))
839 delete_tempfile(&refs->tempfile);
840
841 if (data->own_lock && is_lock_file_locked(&refs->lock)) {
842 packed_refs_unlock(&refs->base);
843 data->own_lock = 0;
844 }
845
846 free(data);
847 transaction->backend_data = NULL;
848 }
849
850 transaction->state = REF_TRANSACTION_CLOSED;
851}
852
853static int packed_transaction_prepare(struct ref_store *ref_store,
854 struct ref_transaction *transaction,
855 struct strbuf *err)
856{
857 struct packed_ref_store *refs = packed_downcast(
858 ref_store,
859 REF_STORE_READ | REF_STORE_WRITE | REF_STORE_ODB,
860 "ref_transaction_prepare");
861 struct packed_transaction_backend_data *data;
862 size_t i;
863 int ret = TRANSACTION_GENERIC_ERROR;
864
865 /*
866 * Note that we *don't* skip transactions with zero updates,
867 * because such a transaction might be executed for the side
868 * effect of ensuring that all of the references are peeled.
869 * If the caller wants to optimize away empty transactions, it
870 * should do so itself.
871 */
872
873 data = xcalloc(1, sizeof(*data));
874 string_list_init(&data->updates, 0);
875
876 transaction->backend_data = data;
877
878 /*
879 * Stick the updates in a string list by refname so that we
880 * can sort them:
881 */
882 for (i = 0; i < transaction->nr; i++) {
883 struct ref_update *update = transaction->updates[i];
884 struct string_list_item *item =
885 string_list_append(&data->updates, update->refname);
886
887 /* Store a pointer to update in item->util: */
888 item->util = update;
889 }
890 string_list_sort(&data->updates);
891
892 if (ref_update_reject_duplicates(&data->updates, err))
893 goto failure;
894
895 if (!is_lock_file_locked(&refs->lock)) {
896 if (packed_refs_lock(ref_store, 0, err))
897 goto failure;
898 data->own_lock = 1;
899 }
900
901 if (write_with_updates(refs, &data->updates, err))
902 goto failure;
903
904 transaction->state = REF_TRANSACTION_PREPARED;
905 return 0;
906
907failure:
908 packed_transaction_cleanup(refs, transaction);
909 return ret;
910}
911
912static int packed_transaction_abort(struct ref_store *ref_store,
913 struct ref_transaction *transaction,
914 struct strbuf *err)
915{
916 struct packed_ref_store *refs = packed_downcast(
917 ref_store,
918 REF_STORE_READ | REF_STORE_WRITE | REF_STORE_ODB,
919 "ref_transaction_abort");
920
921 packed_transaction_cleanup(refs, transaction);
922 return 0;
923}
924
925static int packed_transaction_finish(struct ref_store *ref_store,
926 struct ref_transaction *transaction,
927 struct strbuf *err)
928{
929 struct packed_ref_store *refs = packed_downcast(
930 ref_store,
931 REF_STORE_READ | REF_STORE_WRITE | REF_STORE_ODB,
932 "ref_transaction_finish");
933 int ret = TRANSACTION_GENERIC_ERROR;
934 char *packed_refs_path;
935
936 packed_refs_path = get_locked_file_path(&refs->lock);
937 if (rename_tempfile(&refs->tempfile, packed_refs_path)) {
938 strbuf_addf(err, "error replacing %s: %s",
939 refs->path, strerror(errno));
940 goto cleanup;
941 }
942
943 clear_packed_ref_cache(refs);
944 ret = 0;
945
946cleanup:
947 free(packed_refs_path);
948 packed_transaction_cleanup(refs, transaction);
949 return ret;
950}
951
952static int packed_initial_transaction_commit(struct ref_store *ref_store,
953 struct ref_transaction *transaction,
954 struct strbuf *err)
955{
956 return ref_transaction_commit(transaction, err);
957}
958
959static int packed_delete_refs(struct ref_store *ref_store, const char *msg,
960 struct string_list *refnames, unsigned int flags)
961{
962 struct packed_ref_store *refs =
963 packed_downcast(ref_store, REF_STORE_WRITE, "delete_refs");
964 struct strbuf err = STRBUF_INIT;
965 struct ref_transaction *transaction;
966 struct string_list_item *item;
967 int ret;
968
969 (void)refs; /* We need the check above, but don't use the variable */
970
971 if (!refnames->nr)
972 return 0;
973
974 /*
975 * Since we don't check the references' old_oids, the
976 * individual updates can't fail, so we can pack all of the
977 * updates into a single transaction.
978 */
979
980 transaction = ref_store_transaction_begin(ref_store, &err);
981 if (!transaction)
982 return -1;
983
984 for_each_string_list_item(item, refnames) {
985 if (ref_transaction_delete(transaction, item->string, NULL,
986 flags, msg, &err)) {
987 warning(_("could not delete reference %s: %s"),
988 item->string, err.buf);
989 strbuf_reset(&err);
990 }
991 }
992
993 ret = ref_transaction_commit(transaction, &err);
994
995 if (ret) {
996 if (refnames->nr == 1)
997 error(_("could not delete reference %s: %s"),
998 refnames->items[0].string, err.buf);
999 else
1000 error(_("could not delete references: %s"), err.buf);
1001 }
1002
1003 ref_transaction_free(transaction);
1004 strbuf_release(&err);
1005 return ret;
1006}
1007
1008static int packed_pack_refs(struct ref_store *ref_store, unsigned int flags)
1009{
1010 /*
1011 * Packed refs are already packed. It might be that loose refs
1012 * are packed *into* a packed refs store, but that is done by
1013 * updating the packed references via a transaction.
1014 */
1015 return 0;
1016}
1017
1018static int packed_create_symref(struct ref_store *ref_store,
1019 const char *refname, const char *target,
1020 const char *logmsg)
1021{
1022 die("BUG: packed reference store does not support symrefs");
1023}
1024
1025static int packed_rename_ref(struct ref_store *ref_store,
1026 const char *oldrefname, const char *newrefname,
1027 const char *logmsg)
1028{
1029 die("BUG: packed reference store does not support renaming references");
1030}
1031
1032static struct ref_iterator *packed_reflog_iterator_begin(struct ref_store *ref_store)
1033{
1034 return empty_ref_iterator_begin();
1035}
1036
1037static int packed_for_each_reflog_ent(struct ref_store *ref_store,
1038 const char *refname,
1039 each_reflog_ent_fn fn, void *cb_data)
1040{
1041 return 0;
1042}
1043
1044static int packed_for_each_reflog_ent_reverse(struct ref_store *ref_store,
1045 const char *refname,
1046 each_reflog_ent_fn fn,
1047 void *cb_data)
1048{
1049 return 0;
1050}
1051
1052static int packed_reflog_exists(struct ref_store *ref_store,
1053 const char *refname)
1054{
1055 return 0;
1056}
1057
1058static int packed_create_reflog(struct ref_store *ref_store,
1059 const char *refname, int force_create,
1060 struct strbuf *err)
1061{
1062 die("BUG: packed reference store does not support reflogs");
1063}
1064
1065static int packed_delete_reflog(struct ref_store *ref_store,
1066 const char *refname)
1067{
1068 return 0;
1069}
1070
1071static int packed_reflog_expire(struct ref_store *ref_store,
1072 const char *refname, const unsigned char *sha1,
1073 unsigned int flags,
1074 reflog_expiry_prepare_fn prepare_fn,
1075 reflog_expiry_should_prune_fn should_prune_fn,
1076 reflog_expiry_cleanup_fn cleanup_fn,
1077 void *policy_cb_data)
1078{
1079 return 0;
1080}
1081
1082struct ref_storage_be refs_be_packed = {
1083 NULL,
1084 "packed",
1085 packed_ref_store_create,
1086 packed_init_db,
1087 packed_transaction_prepare,
1088 packed_transaction_finish,
1089 packed_transaction_abort,
1090 packed_initial_transaction_commit,
1091
1092 packed_pack_refs,
1093 packed_peel_ref,
1094 packed_create_symref,
1095 packed_delete_refs,
1096 packed_rename_ref,
1097
1098 packed_ref_iterator_begin,
1099 packed_read_raw_ref,
1100
1101 packed_reflog_iterator_begin,
1102 packed_for_each_reflog_ent,
1103 packed_for_each_reflog_ent_reverse,
1104 packed_reflog_exists,
1105 packed_create_reflog,
1106 packed_delete_reflog,
1107 packed_reflog_expire
1108};