about summary refs log tree commit diff
path: root/third_party/nix/src/libstore/nar-info-disk-cache.cc
#include "nar-info-disk-cache.hh"
#include <sqlite3.h>
#include "globals.hh"
#include "sqlite.hh"
#include "sync.hh"

namespace nix {

static const char* schema = R"sql(

create table if not exists BinaryCaches (
    id        integer primary key autoincrement not null,
    url       text unique not null,
    timestamp integer not null,
    storeDir  text not null,
    wantMassQuery integer not null,
    priority  integer not null
);

create table if not exists NARs (
    cache            integer not null,
    hashPart         text not null,
    namePart         text,
    url              text,
    compression      text,
    fileHash         text,
    fileSize         integer,
    narHash          text,
    narSize          integer,
    refs             text,
    deriver          text,
    sigs             text,
    ca               text,
    timestamp        integer not null,
    present          integer not null,
    primary key (cache, hashPart),
    foreign key (cache) references BinaryCaches(id) on delete cascade
);

create table if not exists LastPurge (
    dummy            text primary key,
    value            integer
);

)sql";

class NarInfoDiskCacheImpl : public NarInfoDiskCache {
 public:
  /* How often to purge expired entries from the cache. */
  const int purgeInterval = 24 * 3600;

  struct Cache {
    int id;
    Path storeDir;
    bool wantMassQuery;
    int priority;
  };

  struct State {
    SQLite db;
    SQLiteStmt insertCache, queryCache, insertNAR, insertMissingNAR, queryNAR,
        purgeCache;
    std::map<std::string, Cache> caches;
  };

  Sync<State> _state;

  NarInfoDiskCacheImpl() {
    auto state(_state.lock());

    Path dbPath = getCacheDir() + "/nix/binary-cache-v6.sqlite";
    createDirs(dirOf(dbPath));

    state->db = SQLite(dbPath);

    if (sqlite3_busy_timeout(state->db, 60 * 60 * 1000) != SQLITE_OK)
      throwSQLiteError(state->db, "setting timeout");

    // We can always reproduce the cache.
    state->db.exec("pragma synchronous = off");
    state->db.exec("pragma main.journal_mode = truncate");

    state->db.exec(schema);

    state->insertCache.create(
        state->db,
        "insert or replace into BinaryCaches(url, timestamp, storeDir, "
        "wantMassQuery, priority) values (?, ?, ?, ?, ?)");

    state->queryCache.create(state->db,
                             "select id, storeDir, wantMassQuery, priority "
                             "from BinaryCaches where url = ?");

    state->insertNAR.create(
        state->db,
        "insert or replace into NARs(cache, hashPart, namePart, url, "
        "compression, fileHash, fileSize, narHash, "
        "narSize, refs, deriver, sigs, ca, timestamp, present) values (?, ?, "
        "?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, 1)");

    state->insertMissingNAR.create(
        state->db,
        "insert or replace into NARs(cache, hashPart, timestamp, present) "
        "values (?, ?, ?, 0)");

    state->queryNAR.create(
        state->db,
        "select present, namePart, url, compression, fileHash, fileSize, "
        "narHash, narSize, refs, deriver, sigs, ca from NARs where cache = ? "
        "and hashPart = ? and ((present = 0 and timestamp > ?) or (present = 1 "
        "and timestamp > ?))");

    /* Periodically purge expired entries from the database. */
    retrySQLite<void>([&]() {
      auto now = time(0);

      SQLiteStmt queryLastPurge(state->db, "select value from LastPurge");
      auto queryLastPurge_(queryLastPurge.use());

      if (!queryLastPurge_.next() ||
          queryLastPurge_.getInt(0) < now - purgeInterval) {
        SQLiteStmt(state->db,
                   "delete from NARs where ((present = 0 and timestamp < ?) or "
                   "(present = 1 and timestamp < ?))")
            .use()(now - settings.ttlNegativeNarInfoCache)(
                now - settings.ttlPositiveNarInfoCache)
            .exec();

        debug("deleted %d entries from the NAR info disk cache",
              sqlite3_changes(state->db));

        SQLiteStmt(
            state->db,
            "insert or replace into LastPurge(dummy, value) values ('', ?)")
            .use()(now)
            .exec();
      }
    });
  }

  Cache& getCache(State& state, const std::string& uri) {
    auto i = state.caches.find(uri);
    if (i == state.caches.end()) abort();
    return i->second;
  }

  void createCache(const std::string& uri, const Path& storeDir,
                   bool wantMassQuery, int priority) override {
    retrySQLite<void>([&]() {
      auto state(_state.lock());

      // FIXME: race

      state->insertCache.use()(uri)(time(0))(storeDir)(wantMassQuery)(priority)
          .exec();
      assert(sqlite3_changes(state->db) == 1);
      state->caches[uri] = Cache{(int)sqlite3_last_insert_rowid(state->db),
                                 storeDir, wantMassQuery, priority};
    });
  }

  bool cacheExists(const std::string& uri, bool& wantMassQuery,
                   int& priority) override {
    return retrySQLite<bool>([&]() {
      auto state(_state.lock());

      auto i = state->caches.find(uri);
      if (i == state->caches.end()) {
        auto queryCache(state->queryCache.use()(uri));
        if (!queryCache.next()) return false;
        state->caches.emplace(
            uri, Cache{(int)queryCache.getInt(0), queryCache.getStr(1),
                       queryCache.getInt(2) != 0, (int)queryCache.getInt(3)});
      }

      auto& cache(getCache(*state, uri));

      wantMassQuery = cache.wantMassQuery;
      priority = cache.priority;

      return true;
    });
  }

  std::pair<Outcome, std::shared_ptr<NarInfo>> lookupNarInfo(
      const std::string& uri, const std::string& hashPart) override {
    return retrySQLite<std::pair<Outcome, std::shared_ptr<NarInfo>>>(
        [&]() -> std::pair<Outcome, std::shared_ptr<NarInfo>> {
          auto state(_state.lock());

          auto& cache(getCache(*state, uri));

          auto now = time(0);

          auto queryNAR(state->queryNAR.use()(cache.id)(hashPart)(
              now - settings.ttlNegativeNarInfoCache)(
              now - settings.ttlPositiveNarInfoCache));

          if (!queryNAR.next()) return {oUnknown, 0};

          if (!queryNAR.getInt(0)) return {oInvalid, 0};

          auto narInfo = make_ref<NarInfo>();

          auto namePart = queryNAR.getStr(1);
          narInfo->path = cache.storeDir + "/" + hashPart +
                          (namePart.empty() ? "" : "-" + namePart);
          narInfo->url = queryNAR.getStr(2);
          narInfo->compression = queryNAR.getStr(3);
          if (!queryNAR.isNull(4)) narInfo->fileHash = Hash(queryNAR.getStr(4));
          narInfo->fileSize = queryNAR.getInt(5);
          narInfo->narHash = Hash(queryNAR.getStr(6));
          narInfo->narSize = queryNAR.getInt(7);
          for (auto& r : tokenizeString<Strings>(queryNAR.getStr(8), " "))
            narInfo->references.insert(cache.storeDir + "/" + r);
          if (!queryNAR.isNull(9))
            narInfo->deriver = cache.storeDir + "/" + queryNAR.getStr(9);
          for (auto& sig : tokenizeString<Strings>(queryNAR.getStr(10), " "))
            narInfo->sigs.insert(sig);
          narInfo->ca = queryNAR.getStr(11);

          return {oValid, narInfo};
        });
  }

  void upsertNarInfo(const std::string& uri, const std::string& hashPart,
                     std::shared_ptr<ValidPathInfo> info) override {
    retrySQLite<void>([&]() {
      auto state(_state.lock());

      auto& cache(getCache(*state, uri));

      if (info) {
        auto narInfo = std::dynamic_pointer_cast<NarInfo>(info);

        assert(hashPart == storePathToHash(info->path));

        state->insertNAR
            .use()(cache.id)(hashPart)(storePathToName(info->path))(
                narInfo ? narInfo->url : "", narInfo != 0)(
                narInfo ? narInfo->compression : "", narInfo != 0)(
                narInfo && narInfo->fileHash ? narInfo->fileHash.to_string()
                                             : "",
                narInfo && narInfo->fileHash)(
                narInfo ? narInfo->fileSize : 0,
                narInfo != 0 && narInfo->fileSize)(info->narHash.to_string())(
                info->narSize)(concatStringsSep(" ", info->shortRefs()))(
                info->deriver != "" ? baseNameOf(info->deriver) : "",
                info->deriver !=
                    "")(concatStringsSep(" ", info->sigs))(info->ca)(time(0))
            .exec();

      } else {
        state->insertMissingNAR.use()(cache.id)(hashPart)(time(0)).exec();
      }
    });
  }
};

ref<NarInfoDiskCache> getNarInfoDiskCache() {
  static ref<NarInfoDiskCache> cache = make_ref<NarInfoDiskCacheImpl>();
  return cache;
}

}  // namespace nix