| #ifndef READ_CACHE_LL_H |
| #define READ_CACHE_LL_H |
| |
| #include "hash-ll.h" |
| #include "hashmap.h" |
| #include "statinfo.h" |
| |
| /* |
| * Basic data structures for the directory cache |
| */ |
| |
| #define CACHE_SIGNATURE 0x44495243 /* "DIRC" */ |
| struct cache_header { |
| uint32_t hdr_signature; |
| uint32_t hdr_version; |
| uint32_t hdr_entries; |
| }; |
| |
| #define INDEX_FORMAT_LB 2 |
| #define INDEX_FORMAT_UB 4 |
| |
| struct cache_entry { |
| struct hashmap_entry ent; |
| struct stat_data ce_stat_data; |
| unsigned int ce_mode; |
| unsigned int ce_flags; |
| unsigned int mem_pool_allocated; |
| unsigned int ce_namelen; |
| unsigned int index; /* for link extension */ |
| struct object_id oid; |
| char name[FLEX_ARRAY]; /* more */ |
| }; |
| |
| #define CE_STAGEMASK (0x3000) |
| #define CE_EXTENDED (0x4000) |
| #define CE_VALID (0x8000) |
| #define CE_STAGESHIFT 12 |
| |
| /* |
| * Range 0xFFFF0FFF in ce_flags is divided into |
| * two parts: in-memory flags and on-disk ones. |
| * Flags in CE_EXTENDED_FLAGS will get saved on-disk |
| * if you want to save a new flag, add it in |
| * CE_EXTENDED_FLAGS |
| * |
| * In-memory only flags |
| */ |
| #define CE_UPDATE (1 << 16) |
| #define CE_REMOVE (1 << 17) |
| #define CE_UPTODATE (1 << 18) |
| #define CE_ADDED (1 << 19) |
| |
| #define CE_HASHED (1 << 20) |
| #define CE_FSMONITOR_VALID (1 << 21) |
| #define CE_WT_REMOVE (1 << 22) /* remove in work directory */ |
| #define CE_CONFLICTED (1 << 23) |
| |
| #define CE_UNPACKED (1 << 24) |
| #define CE_NEW_SKIP_WORKTREE (1 << 25) |
| |
| /* used to temporarily mark paths matched by pathspecs */ |
| #define CE_MATCHED (1 << 26) |
| |
| #define CE_UPDATE_IN_BASE (1 << 27) |
| #define CE_STRIP_NAME (1 << 28) |
| |
| /* |
| * Extended on-disk flags |
| */ |
| #define CE_INTENT_TO_ADD (1 << 29) |
| #define CE_SKIP_WORKTREE (1 << 30) |
| /* CE_EXTENDED2 is for future extension */ |
| #define CE_EXTENDED2 (1U << 31) |
| |
| #define CE_EXTENDED_FLAGS (CE_INTENT_TO_ADD | CE_SKIP_WORKTREE) |
| |
| /* |
| * Safeguard to avoid saving wrong flags: |
| * - CE_EXTENDED2 won't get saved until its semantic is known |
| * - Bits in 0x0000FFFF have been saved in ce_flags already |
| * - Bits in 0x003F0000 are currently in-memory flags |
| */ |
| #if CE_EXTENDED_FLAGS & 0x803FFFFF |
| #error "CE_EXTENDED_FLAGS out of range" |
| #endif |
| |
| /* Forward structure decls */ |
| struct pathspec; |
| struct tree; |
| |
| /* |
| * Copy the sha1 and stat state of a cache entry from one to |
| * another. But we never change the name, or the hash state! |
| */ |
| static inline void copy_cache_entry(struct cache_entry *dst, |
| const struct cache_entry *src) |
| { |
| unsigned int state = dst->ce_flags & CE_HASHED; |
| int mem_pool_allocated = dst->mem_pool_allocated; |
| |
| /* Don't copy hash chain and name */ |
| memcpy(&dst->ce_stat_data, &src->ce_stat_data, |
| offsetof(struct cache_entry, name) - |
| offsetof(struct cache_entry, ce_stat_data)); |
| |
| /* Restore the hash state */ |
| dst->ce_flags = (dst->ce_flags & ~CE_HASHED) | state; |
| |
| /* Restore the mem_pool_allocated flag */ |
| dst->mem_pool_allocated = mem_pool_allocated; |
| } |
| |
| static inline unsigned create_ce_flags(unsigned stage) |
| { |
| return (stage << CE_STAGESHIFT); |
| } |
| |
| #define ce_namelen(ce) ((ce)->ce_namelen) |
| #define ce_size(ce) cache_entry_size(ce_namelen(ce)) |
| #define ce_stage(ce) ((CE_STAGEMASK & (ce)->ce_flags) >> CE_STAGESHIFT) |
| #define ce_uptodate(ce) ((ce)->ce_flags & CE_UPTODATE) |
| #define ce_skip_worktree(ce) ((ce)->ce_flags & CE_SKIP_WORKTREE) |
| #define ce_mark_uptodate(ce) ((ce)->ce_flags |= CE_UPTODATE) |
| #define ce_intent_to_add(ce) ((ce)->ce_flags & CE_INTENT_TO_ADD) |
| |
| #define cache_entry_size(len) (offsetof(struct cache_entry,name) + (len) + 1) |
| |
| #define SOMETHING_CHANGED (1 << 0) /* unclassified changes go here */ |
| #define CE_ENTRY_CHANGED (1 << 1) |
| #define CE_ENTRY_REMOVED (1 << 2) |
| #define CE_ENTRY_ADDED (1 << 3) |
| #define RESOLVE_UNDO_CHANGED (1 << 4) |
| #define CACHE_TREE_CHANGED (1 << 5) |
| #define SPLIT_INDEX_ORDERED (1 << 6) |
| #define UNTRACKED_CHANGED (1 << 7) |
| #define FSMONITOR_CHANGED (1 << 8) |
| |
| struct split_index; |
| struct untracked_cache; |
| struct progress; |
| struct pattern_list; |
| |
| enum sparse_index_mode { |
| /* |
| * There are no sparse directories in the index at all. |
| * |
| * Repositories that don't use cone-mode sparse-checkout will |
| * always have their indexes in this mode. |
| */ |
| INDEX_EXPANDED = 0, |
| |
| /* |
| * The index has already been collapsed to sparse directories |
| * whereever possible. |
| */ |
| INDEX_COLLAPSED, |
| |
| /* |
| * The sparse directories that exist are outside the |
| * sparse-checkout boundary, but it is possible that some file |
| * entries could collapse to sparse directory entries. |
| */ |
| INDEX_PARTIALLY_SPARSE, |
| }; |
| |
| struct index_state { |
| struct cache_entry **cache; |
| unsigned int version; |
| unsigned int cache_nr, cache_alloc, cache_changed; |
| struct string_list *resolve_undo; |
| struct cache_tree *cache_tree; |
| struct split_index *split_index; |
| struct cache_time timestamp; |
| unsigned name_hash_initialized : 1, |
| initialized : 1, |
| drop_cache_tree : 1, |
| updated_workdir : 1, |
| updated_skipworktree : 1, |
| fsmonitor_has_run_once : 1; |
| enum sparse_index_mode sparse_index; |
| struct hashmap name_hash; |
| struct hashmap dir_hash; |
| struct object_id oid; |
| struct untracked_cache *untracked; |
| char *fsmonitor_last_update; |
| struct ewah_bitmap *fsmonitor_dirty; |
| struct mem_pool *ce_mem_pool; |
| struct progress *progress; |
| struct repository *repo; |
| struct pattern_list *sparse_checkout_patterns; |
| }; |
| |
| /** |
| * A "struct index_state istate" must be initialized with |
| * INDEX_STATE_INIT or the corresponding index_state_init(). |
| * |
| * If the variable won't be used again, use release_index() to free() |
| * its resources. If it needs to be used again use discard_index(), |
| * which does the same thing, but will use use index_state_init() at |
| * the end. The discard_index() will use its own "istate->repo" as the |
| * "r" argument to index_state_init() in that case. |
| */ |
| #define INDEX_STATE_INIT(r) { \ |
| .repo = (r), \ |
| } |
| void index_state_init(struct index_state *istate, struct repository *r); |
| void release_index(struct index_state *istate); |
| |
| /* Cache entry creation and cleanup */ |
| |
| /* |
| * Create cache_entry intended for use in the specified index. Caller |
| * is responsible for discarding the cache_entry with |
| * `discard_cache_entry`. |
| */ |
| struct cache_entry *make_cache_entry(struct index_state *istate, |
| unsigned int mode, |
| const struct object_id *oid, |
| const char *path, |
| int stage, |
| unsigned int refresh_options); |
| |
| struct cache_entry *make_empty_cache_entry(struct index_state *istate, |
| size_t name_len); |
| |
| /* |
| * Create a cache_entry that is not intended to be added to an index. If |
| * `ce_mem_pool` is not NULL, the entry is allocated within the given memory |
| * pool. Caller is responsible for discarding "loose" entries with |
| * `discard_cache_entry()` and the memory pool with |
| * `mem_pool_discard(ce_mem_pool, should_validate_cache_entries())`. |
| */ |
| struct cache_entry *make_transient_cache_entry(unsigned int mode, |
| const struct object_id *oid, |
| const char *path, |
| int stage, |
| struct mem_pool *ce_mem_pool); |
| |
| struct cache_entry *make_empty_transient_cache_entry(size_t len, |
| struct mem_pool *ce_mem_pool); |
| |
| /* |
| * Discard cache entry. |
| */ |
| void discard_cache_entry(struct cache_entry *ce); |
| |
| /* |
| * Check configuration if we should perform extra validation on cache |
| * entries. |
| */ |
| int should_validate_cache_entries(void); |
| |
| /* |
| * Duplicate a cache_entry. Allocate memory for the new entry from a |
| * memory_pool. Takes into account cache_entry fields that are meant |
| * for managing the underlying memory allocation of the cache_entry. |
| */ |
| struct cache_entry *dup_cache_entry(const struct cache_entry *ce, struct index_state *istate); |
| |
| /* |
| * Validate the cache entries in the index. This is an internal |
| * consistency check that the cache_entry structs are allocated from |
| * the expected memory pool. |
| */ |
| void validate_cache_entries(const struct index_state *istate); |
| |
| /* |
| * Bulk prefetch all missing cache entries that are not GITLINKs and that match |
| * the given predicate. This function should only be called if |
| * repo_has_promisor_remote() returns true. |
| */ |
| typedef int (*must_prefetch_predicate)(const struct cache_entry *); |
| void prefetch_cache_entries(const struct index_state *istate, |
| must_prefetch_predicate must_prefetch); |
| |
| /* Initialize and use the cache information */ |
| struct lock_file; |
| int do_read_index(struct index_state *istate, const char *path, |
| int must_exist); /* for testting only! */ |
| int read_index_from(struct index_state *, const char *path, |
| const char *gitdir); |
| int is_index_unborn(struct index_state *); |
| |
| /* For use with `write_locked_index()`. */ |
| #define COMMIT_LOCK (1 << 0) |
| #define SKIP_IF_UNCHANGED (1 << 1) |
| |
| /* |
| * Write the index while holding an already-taken lock. Close the lock, |
| * and if `COMMIT_LOCK` is given, commit it. |
| * |
| * Unless a split index is in use, write the index into the lockfile. |
| * |
| * With a split index, write the shared index to a temporary file, |
| * adjust its permissions and rename it into place, then write the |
| * split index to the lockfile. If the temporary file for the shared |
| * index cannot be created, fall back to the behavior described in |
| * the previous paragraph. |
| * |
| * With `COMMIT_LOCK`, the lock is always committed or rolled back. |
| * Without it, the lock is closed, but neither committed nor rolled |
| * back. |
| * |
| * If `SKIP_IF_UNCHANGED` is given and the index is unchanged, nothing |
| * is written (and the lock is rolled back if `COMMIT_LOCK` is given). |
| */ |
| int write_locked_index(struct index_state *, struct lock_file *lock, unsigned flags); |
| |
| void discard_index(struct index_state *); |
| void move_index_extensions(struct index_state *dst, struct index_state *src); |
| int unmerged_index(const struct index_state *); |
| |
| /** |
| * Returns 1 if istate differs from tree, 0 otherwise. If tree is NULL, |
| * compares istate to HEAD. If tree is NULL and on an unborn branch, |
| * returns 1 if there are entries in istate, 0 otherwise. If an strbuf is |
| * provided, the space-separated list of files that differ will be appended |
| * to it. |
| */ |
| int repo_index_has_changes(struct repository *repo, |
| struct tree *tree, |
| struct strbuf *sb); |
| |
| int verify_path(const char *path, unsigned mode); |
| int strcmp_offset(const char *s1, const char *s2, size_t *first_change); |
| |
| /* |
| * Searches for an entry defined by name and namelen in the given index. |
| * If the return value is positive (including 0) it is the position of an |
| * exact match. If the return value is negative, the negated value minus 1 |
| * is the position where the entry would be inserted. |
| * Example: The current index consists of these files and its stages: |
| * |
| * b#0, d#0, f#1, f#3 |
| * |
| * index_name_pos(&index, "a", 1) -> -1 |
| * index_name_pos(&index, "b", 1) -> 0 |
| * index_name_pos(&index, "c", 1) -> -2 |
| * index_name_pos(&index, "d", 1) -> 1 |
| * index_name_pos(&index, "e", 1) -> -3 |
| * index_name_pos(&index, "f", 1) -> -3 |
| * index_name_pos(&index, "g", 1) -> -5 |
| */ |
| int index_name_pos(struct index_state *, const char *name, int namelen); |
| |
| /* |
| * Like index_name_pos, returns the position of an entry of the given name in |
| * the index if one exists, otherwise returns a negative value where the negated |
| * value minus 1 is the position where the index entry would be inserted. Unlike |
| * index_name_pos, however, a sparse index is not expanded to find an entry |
| * inside a sparse directory. |
| */ |
| int index_name_pos_sparse(struct index_state *, const char *name, int namelen); |
| |
| /* |
| * Determines whether an entry with the given name exists within the |
| * given index. The return value is 1 if an exact match is found, otherwise |
| * it is 0. Note that, unlike index_name_pos, this function does not expand |
| * the index if it is sparse. If an item exists within the full index but it |
| * is contained within a sparse directory (and not in the sparse index), 0 is |
| * returned. |
| */ |
| int index_entry_exists(struct index_state *, const char *name, int namelen); |
| |
| /* |
| * Some functions return the negative complement of an insert position when a |
| * precise match was not found but a position was found where the entry would |
| * need to be inserted. This helper protects that logic from any integer |
| * underflow. |
| */ |
| static inline int index_pos_to_insert_pos(uintmax_t pos) |
| { |
| if (pos > INT_MAX) |
| die("overflow: -1 - %"PRIuMAX, pos); |
| return -1 - (int)pos; |
| } |
| |
| #define ADD_CACHE_OK_TO_ADD 1 /* Ok to add */ |
| #define ADD_CACHE_OK_TO_REPLACE 2 /* Ok to replace file/directory */ |
| #define ADD_CACHE_SKIP_DFCHECK 4 /* Ok to skip DF conflict checks */ |
| #define ADD_CACHE_JUST_APPEND 8 /* Append only */ |
| #define ADD_CACHE_NEW_ONLY 16 /* Do not replace existing ones */ |
| #define ADD_CACHE_KEEP_CACHE_TREE 32 /* Do not invalidate cache-tree */ |
| #define ADD_CACHE_RENORMALIZE 64 /* Pass along HASH_RENORMALIZE */ |
| int add_index_entry(struct index_state *, struct cache_entry *ce, int option); |
| void rename_index_entry_at(struct index_state *, int pos, const char *new_name); |
| |
| /* Remove entry, return true if there are more entries to go. */ |
| int remove_index_entry_at(struct index_state *, int pos); |
| |
| void remove_marked_cache_entries(struct index_state *istate, int invalidate); |
| int remove_file_from_index(struct index_state *, const char *path); |
| #define ADD_CACHE_VERBOSE 1 |
| #define ADD_CACHE_PRETEND 2 |
| #define ADD_CACHE_IGNORE_ERRORS 4 |
| #define ADD_CACHE_IGNORE_REMOVAL 8 |
| #define ADD_CACHE_INTENT 16 |
| /* |
| * These two are used to add the contents of the file at path |
| * to the index, marking the working tree up-to-date by storing |
| * the cached stat info in the resulting cache entry. A caller |
| * that has already run lstat(2) on the path can call |
| * add_to_index(), and all others can call add_file_to_index(); |
| * the latter will do necessary lstat(2) internally before |
| * calling the former. |
| */ |
| int add_to_index(struct index_state *, const char *path, struct stat *, int flags); |
| int add_file_to_index(struct index_state *, const char *path, int flags); |
| |
| int chmod_index_entry(struct index_state *, struct cache_entry *ce, char flip); |
| int ce_same_name(const struct cache_entry *a, const struct cache_entry *b); |
| void set_object_name_for_intent_to_add_entry(struct cache_entry *ce); |
| int index_name_is_other(struct index_state *, const char *, int); |
| void *read_blob_data_from_index(struct index_state *, const char *, unsigned long *); |
| |
| /* do stat comparison even if CE_VALID is true */ |
| #define CE_MATCH_IGNORE_VALID 01 |
| /* do not check the contents but report dirty on racily-clean entries */ |
| #define CE_MATCH_RACY_IS_DIRTY 02 |
| /* do stat comparison even if CE_SKIP_WORKTREE is true */ |
| #define CE_MATCH_IGNORE_SKIP_WORKTREE 04 |
| /* ignore non-existent files during stat update */ |
| #define CE_MATCH_IGNORE_MISSING 0x08 |
| /* enable stat refresh */ |
| #define CE_MATCH_REFRESH 0x10 |
| /* don't refresh_fsmonitor state or do stat comparison even if CE_FSMONITOR_VALID is true */ |
| #define CE_MATCH_IGNORE_FSMONITOR 0X20 |
| int is_racy_timestamp(const struct index_state *istate, |
| const struct cache_entry *ce); |
| int has_racy_timestamp(struct index_state *istate); |
| int ie_match_stat(struct index_state *, const struct cache_entry *, struct stat *, unsigned int); |
| int ie_modified(struct index_state *, const struct cache_entry *, struct stat *, unsigned int); |
| |
| int match_stat_data_racy(const struct index_state *istate, |
| const struct stat_data *sd, struct stat *st); |
| |
| void fill_stat_cache_info(struct index_state *istate, struct cache_entry *ce, struct stat *st); |
| |
| /* |
| * Fill members of st by members of sd enough to convince match_stat() |
| * to consider that they match. It should be usable as a replacement |
| * for lstat() for a tracked path that is known to be up-to-date via |
| * some out-of-line means (like fsmonitor). |
| */ |
| int fake_lstat(const struct cache_entry *ce, struct stat *st); |
| |
| #define REFRESH_REALLY (1 << 0) /* ignore_valid */ |
| #define REFRESH_UNMERGED (1 << 1) /* allow unmerged */ |
| #define REFRESH_QUIET (1 << 2) /* be quiet about it */ |
| #define REFRESH_IGNORE_MISSING (1 << 3) /* ignore non-existent */ |
| #define REFRESH_IGNORE_SUBMODULES (1 << 4) /* ignore submodules */ |
| #define REFRESH_IN_PORCELAIN (1 << 5) /* user friendly output, not "needs update" */ |
| #define REFRESH_PROGRESS (1 << 6) /* show progress bar if stderr is tty */ |
| #define REFRESH_IGNORE_SKIP_WORKTREE (1 << 7) /* ignore skip_worktree entries */ |
| int refresh_index(struct index_state *, unsigned int flags, const struct pathspec *pathspec, char *seen, const char *header_msg); |
| /* |
| * Refresh the index and write it to disk. |
| * |
| * 'refresh_flags' is passed directly to 'refresh_index()', while |
| * 'COMMIT_LOCK | write_flags' is passed to 'write_locked_index()', so |
| * the lockfile is always either committed or rolled back. |
| * |
| * If 'gentle' is passed, errors locking the index are ignored. |
| * |
| * Return 1 if refreshing the index returns an error, -1 if writing |
| * the index to disk fails, 0 on success. |
| * |
| * Note that if refreshing the index returns an error, we still write |
| * out the index (unless locking fails). |
| */ |
| int repo_refresh_and_write_index(struct repository*, unsigned int refresh_flags, unsigned int write_flags, int gentle, const struct pathspec *, char *seen, const char *header_msg); |
| |
| struct cache_entry *refresh_cache_entry(struct index_state *, struct cache_entry *, unsigned int); |
| |
| void set_alternate_index_output(const char *); |
| |
| extern int verify_index_checksum; |
| extern int verify_ce_order; |
| |
| int cmp_cache_name_compare(const void *a_, const void *b_); |
| |
| int add_files_to_cache(struct repository *repo, const char *prefix, |
| const struct pathspec *pathspec, char *ps_matched, |
| int include_sparse, int flags); |
| |
| void overlay_tree_on_index(struct index_state *istate, |
| const char *tree_name, const char *prefix); |
| |
| #endif /* READ_CACHE_LL_H */ |