diff options
author | Vincent Ambo <mail@tazj.in> | 2021-09-21T10·03+0300 |
---|---|---|
committer | Vincent Ambo <mail@tazj.in> | 2021-09-21T11·29+0300 |
commit | 43b1791ec601732ac31195df96781a848360a9ac (patch) | |
tree | daae8d638343295d2f1f7da955e556ef4c958864 /third_party/git/t/helper/test-lazy-init-name-hash.c | |
parent | 2d8e7dc9d9c38127ec4ebd13aee8e8f586a43318 (diff) |
chore(3p/git): Unvendor git and track patches instead r/2903
This was vendored a long time ago under the expectation that keeping it in sync with cgit would be easier this way, but it has proven not to be a big issue. On the other hand, a vendored copy of git is an annoying maintenance burden. It is much easier to rebase the single (dottime) patch that we have. This removes the vendored copy of git and instead passes the git source code to cgit via `pkgs.srcOnly`, which includes the applied patch so that cgit can continue rendering dottime. Change-Id: If31f62dea7ce688fd1b9050204e9378019775f2b
Diffstat (limited to 'third_party/git/t/helper/test-lazy-init-name-hash.c')
-rw-r--r-- | third_party/git/t/helper/test-lazy-init-name-hash.c | 261 |
1 files changed, 0 insertions, 261 deletions
diff --git a/third_party/git/t/helper/test-lazy-init-name-hash.c b/third_party/git/t/helper/test-lazy-init-name-hash.c deleted file mode 100644 index cd1b4c9736ef..000000000000 --- a/third_party/git/t/helper/test-lazy-init-name-hash.c +++ /dev/null @@ -1,261 +0,0 @@ -#include "test-tool.h" -#include "cache.h" -#include "parse-options.h" - -static int single; -static int multi; -static int count = 1; -static int dump; -static int perf; -static int analyze; -static int analyze_step; - -/* - * Dump the contents of the "dir" and "name" hash tables to stdout. - * If you sort the result, you can compare it with the other type - * mode and verify that both single and multi produce the same set. - */ -static void dump_run(void) -{ - struct hashmap_iter iter_dir; - struct hashmap_iter iter_cache; - - /* Stolen from name-hash.c */ - struct dir_entry { - struct hashmap_entry ent; - struct dir_entry *parent; - int nr; - unsigned int namelen; - char name[FLEX_ARRAY]; - }; - - struct dir_entry *dir; - struct cache_entry *ce; - - read_cache(); - if (single) { - test_lazy_init_name_hash(&the_index, 0); - } else { - int nr_threads_used = test_lazy_init_name_hash(&the_index, 1); - if (!nr_threads_used) - die("non-threaded code path used"); - } - - hashmap_for_each_entry(&the_index.dir_hash, &iter_dir, dir, - ent /* member name */) - printf("dir %08x %7d %s\n", dir->ent.hash, dir->nr, dir->name); - - hashmap_for_each_entry(&the_index.name_hash, &iter_cache, ce, - ent /* member name */) - printf("name %08x %s\n", ce->ent.hash, ce->name); - - discard_cache(); -} - -/* - * Run the single or multi threaded version "count" times and - * report on the time taken. - */ -static uint64_t time_runs(int try_threaded) -{ - uint64_t t0, t1, t2; - uint64_t sum = 0; - uint64_t avg; - int nr_threads_used; - int i; - - for (i = 0; i < count; i++) { - t0 = getnanotime(); - read_cache(); - t1 = getnanotime(); - nr_threads_used = test_lazy_init_name_hash(&the_index, try_threaded); - t2 = getnanotime(); - - sum += (t2 - t1); - - if (try_threaded && !nr_threads_used) - die("non-threaded code path used"); - - if (nr_threads_used) - printf("%f %f %d multi %d\n", - ((double)(t1 - t0))/1000000000, - ((double)(t2 - t1))/1000000000, - the_index.cache_nr, - nr_threads_used); - else - printf("%f %f %d single\n", - ((double)(t1 - t0))/1000000000, - ((double)(t2 - t1))/1000000000, - the_index.cache_nr); - fflush(stdout); - - discard_cache(); - } - - avg = sum / count; - if (count > 1) - printf("avg %f %s\n", - (double)avg/1000000000, - (try_threaded) ? "multi" : "single"); - - return avg; -} - -/* - * Try a series of runs varying the "istate->cache_nr" and - * try to find a good value for the multi-threaded criteria. - */ -static void analyze_run(void) -{ - uint64_t t1s, t1m, t2s, t2m; - int cache_nr_limit; - int nr_threads_used = 0; - int i; - int nr; - - read_cache(); - cache_nr_limit = the_index.cache_nr; - discard_cache(); - - nr = analyze; - while (1) { - uint64_t sum_single = 0; - uint64_t sum_multi = 0; - uint64_t avg_single; - uint64_t avg_multi; - - if (nr > cache_nr_limit) - nr = cache_nr_limit; - - for (i = 0; i < count; i++) { - read_cache(); - the_index.cache_nr = nr; /* cheap truncate of index */ - t1s = getnanotime(); - test_lazy_init_name_hash(&the_index, 0); - t2s = getnanotime(); - sum_single += (t2s - t1s); - the_index.cache_nr = cache_nr_limit; - discard_cache(); - - read_cache(); - the_index.cache_nr = nr; /* cheap truncate of index */ - t1m = getnanotime(); - nr_threads_used = test_lazy_init_name_hash(&the_index, 1); - t2m = getnanotime(); - sum_multi += (t2m - t1m); - the_index.cache_nr = cache_nr_limit; - discard_cache(); - - if (!nr_threads_used) - printf(" [size %8d] [single %f] non-threaded code path used\n", - nr, ((double)(t2s - t1s))/1000000000); - else - printf(" [size %8d] [single %f] %c [multi %f %d]\n", - nr, - ((double)(t2s - t1s))/1000000000, - (((t2s - t1s) < (t2m - t1m)) ? '<' : '>'), - ((double)(t2m - t1m))/1000000000, - nr_threads_used); - fflush(stdout); - } - if (count > 1) { - avg_single = sum_single / count; - avg_multi = sum_multi / count; - if (!nr_threads_used) - printf("avg [size %8d] [single %f]\n", - nr, - (double)avg_single/1000000000); - else - printf("avg [size %8d] [single %f] %c [multi %f %d]\n", - nr, - (double)avg_single/1000000000, - (avg_single < avg_multi ? '<' : '>'), - (double)avg_multi/1000000000, - nr_threads_used); - fflush(stdout); - } - - if (nr >= cache_nr_limit) - return; - nr += analyze_step; - } -} - -int cmd__lazy_init_name_hash(int argc, const char **argv) -{ - const char *usage[] = { - "test-tool lazy-init-name-hash -d (-s | -m)", - "test-tool lazy-init-name-hash -p [-c c]", - "test-tool lazy-init-name-hash -a a [--step s] [-c c]", - "test-tool lazy-init-name-hash (-s | -m) [-c c]", - "test-tool lazy-init-name-hash -s -m [-c c]", - NULL - }; - struct option options[] = { - OPT_BOOL('s', "single", &single, "run single-threaded code"), - OPT_BOOL('m', "multi", &multi, "run multi-threaded code"), - OPT_INTEGER('c', "count", &count, "number of passes"), - OPT_BOOL('d', "dump", &dump, "dump hash tables"), - OPT_BOOL('p', "perf", &perf, "compare single vs multi"), - OPT_INTEGER('a', "analyze", &analyze, "analyze different multi sizes"), - OPT_INTEGER(0, "step", &analyze_step, "analyze step factor"), - OPT_END(), - }; - const char *prefix; - uint64_t avg_single, avg_multi; - - prefix = setup_git_directory(); - - argc = parse_options(argc, argv, prefix, options, usage, 0); - - /* - * istate->dir_hash is only created when ignore_case is set. - */ - ignore_case = 1; - - if (dump) { - if (perf || analyze > 0) - die("cannot combine dump, perf, or analyze"); - if (count > 1) - die("count not valid with dump"); - if (single && multi) - die("cannot use both single and multi with dump"); - if (!single && !multi) - die("dump requires either single or multi"); - dump_run(); - return 0; - } - - if (perf) { - if (analyze > 0) - die("cannot combine dump, perf, or analyze"); - if (single || multi) - die("cannot use single or multi with perf"); - avg_single = time_runs(0); - avg_multi = time_runs(1); - if (avg_multi > avg_single) - die("multi is slower"); - return 0; - } - - if (analyze) { - if (analyze < 500) - die("analyze must be at least 500"); - if (!analyze_step) - analyze_step = analyze; - if (single || multi) - die("cannot use single or multi with analyze"); - analyze_run(); - return 0; - } - - if (!single && !multi) - die("require either -s or -m or both"); - - if (single) - time_runs(0); - if (multi) - time_runs(1); - - return 0; -} |