//! Import from a real filesystem.
use futures::stream::BoxStream;
use futures::StreamExt;
use std::fs::FileType;
use std::os::unix::ffi::OsStringExt;
use std::os::unix::fs::MetadataExt;
use std::os::unix::fs::PermissionsExt;
use tokio::io::BufReader;
use tokio_util::io::InspectReader;
use tracing::info_span;
use tracing::instrument;
use tracing::Instrument;
use tracing::Span;
use tracing_indicatif::span_ext::IndicatifSpanExt;
use walkdir::DirEntry;
use walkdir::WalkDir;
use crate::blobservice::BlobService;
use crate::directoryservice::DirectoryService;
use crate::refscan::{ReferenceReader, ReferenceScanner};
use crate::{B3Digest, Node};
use super::ingest_entries;
use super::IngestionEntry;
use super::IngestionError;
/// Ingests the contents at a given path into the tvix store, interacting with a [BlobService] and
/// [DirectoryService]. It returns the root node or an error.
///
/// It does not follow symlinks at the root, they will be ingested as actual symlinks.
///
/// This function will walk the filesystem using `walkdir` and will consume
/// `O(#number of entries)` space.
#[instrument(
skip(blob_service, directory_service, reference_scanner),
fields(path),
err
)]
pub async fn ingest_path<BS, DS, P, P2>(
blob_service: BS,
directory_service: DS,
path: P,
reference_scanner: Option<&ReferenceScanner<P2>>,
) -> Result<Node, IngestionError<Error>>
where
P: AsRef<std::path::Path> + std::fmt::Debug,
BS: BlobService + Clone,
DS: DirectoryService,
P2: AsRef<[u8]> + Send + Sync,
{
let span = Span::current();
let iter = WalkDir::new(path.as_ref())
.follow_links(false)
.follow_root_links(false)
.contents_first(true)
.into_iter();
let entries =
dir_entries_to_ingestion_stream(blob_service, iter, path.as_ref(), reference_scanner);
ingest_entries(
directory_service,
entries.inspect({
let span = span.clone();
move |e| {
if e.is_ok() {
span.pb_inc(1)
}
}
}),
)
.await
}
/// Converts an iterator of [walkdir::DirEntry]s into a stream of ingestion entries.
/// This can then be fed into [ingest_entries] to ingest all the entries into the castore.
///
/// The produced stream is buffered, so uploads can happen concurrently.
///
/// The root is the [Path] in the filesystem that is being ingested into the castore.
pub fn dir_entries_to_ingestion_stream<'a, BS, I, P>(
blob_service: BS,
iter: I,
root: &'a std::path::Path,
reference_scanner: Option<&'a ReferenceScanner<P>>,
) -> BoxStream<'a, Result<IngestionEntry, Error>>
where
BS: BlobService + Clone + 'a,
I: Iterator<Item = Result<DirEntry, walkdir::Error>> + Send + 'a,
P: AsRef<[u8]> + Send + Sync,
{
let prefix = root.parent().unwrap_or_else(|| std::path::Path::new(""));
Box::pin(
futures::stream::iter(iter)
.map(move |x| {
let blob_service = blob_service.clone();
async move {
match x {
Ok(dir_entry) => {
dir_entry_to_ingestion_entry(
blob_service,
&dir_entry,
prefix,
reference_scanner,
)
.await
}
Err(e) => Err(Error::Stat(
prefix.to_path_buf(),
e.into_io_error().expect("walkdir err must be some"),
)),
}
}
})
.buffered(50),
)
}
/// Converts a [walkdir::DirEntry] into an [IngestionEntry], uploading blobs to the
/// provided [BlobService].
///
/// The prefix path is stripped from the path of each entry. This is usually the parent path
/// of the path being ingested so that the last element of the stream only has one component.
pub async fn dir_entry_to_ingestion_entry<BS, P>(
blob_service: BS,
entry: &DirEntry,
prefix: &std::path::Path,
reference_scanner: Option<&ReferenceScanner<P>>,
) -> Result<IngestionEntry, Error>
where
BS: BlobService,
P: AsRef<[u8]>,
{
let file_type = entry.file_type();
let fs_path = entry
.path()
.strip_prefix(prefix)
.expect("Tvix bug: failed to strip root path prefix");
// convert to castore PathBuf
let path = crate::path::PathBuf::from_host_path(fs_path, false)
.unwrap_or_else(|e| panic!("Tvix bug: walkdir direntry cannot be parsed: {}", e));
if file_type.is_dir() {
Ok(IngestionEntry::Dir { path })
} else if file_type.is_symlink() {
let target = std::fs::read_link(entry.path())
.map_err(|e| Error::Stat(entry.path().to_path_buf(), e))?
.into_os_string()
.into_vec();
if let Some(reference_scanner) = &reference_scanner {
reference_scanner.scan(&target);
}
Ok(IngestionEntry::Symlink { path, target })
} else if file_type.is_file() {
let metadata = entry
.metadata()
.map_err(|e| Error::Stat(entry.path().to_path_buf(), e.into()))?;
let digest = upload_blob(blob_service, entry.path().to_path_buf(), reference_scanner)
.instrument({
let span = info_span!("upload_blob", "indicatif.pb_show" = tracing::field::Empty);
span.pb_set_message(&format!("Uploading blob for {:?}", fs_path));
span.pb_set_style(&tvix_tracing::PB_TRANSFER_STYLE);
span
})
.await?;
Ok(IngestionEntry::Regular {
path,
size: metadata.size(),
// If it's executable by the user, it'll become executable.
// This matches nix's dump() function behaviour.
executable: metadata.permissions().mode() & 64 != 0,
digest,
})
} else {
return Err(Error::FileType(fs_path.to_path_buf(), file_type));
}
}
/// Uploads the file at the provided [Path] the the [BlobService].
#[instrument(skip(blob_service, reference_scanner), fields(path), err)]
async fn upload_blob<BS, P>(
blob_service: BS,
path: impl AsRef<std::path::Path>,
reference_scanner: Option<&ReferenceScanner<P>>,
) -> Result<B3Digest, Error>
where
BS: BlobService,
P: AsRef<[u8]>,
{
let span = Span::current();
span.pb_start();
let file = tokio::fs::File::open(path.as_ref())
.await
.map_err(|e| Error::BlobRead(path.as_ref().to_path_buf(), e))?;
let metadata = file
.metadata()
.await
.map_err(|e| Error::Stat(path.as_ref().to_path_buf(), e))?;
span.pb_set_length(metadata.len());
let reader = InspectReader::new(file, |d| {
span.pb_inc(d.len() as u64);
});
let mut writer = blob_service.open_write().await;
if let Some(reference_scanner) = reference_scanner {
let mut reader = ReferenceReader::new(reference_scanner, BufReader::new(reader));
tokio::io::copy(&mut reader, &mut writer)
.await
.map_err(|e| Error::BlobRead(path.as_ref().to_path_buf(), e))?;
} else {
tokio::io::copy(&mut BufReader::new(reader), &mut writer)
.await
.map_err(|e| Error::BlobRead(path.as_ref().to_path_buf(), e))?;
}
let digest = writer
.close()
.await
.map_err(|e| Error::BlobFinalize(path.as_ref().to_path_buf(), e))?;
Ok(digest)
}
#[derive(Debug, thiserror::Error)]
pub enum Error {
#[error("unsupported file type at {0}: {1:?}")]
FileType(std::path::PathBuf, FileType),
#[error("unable to stat {0}: {1}")]
Stat(std::path::PathBuf, std::io::Error),
#[error("unable to open {0}: {1}")]
Open(std::path::PathBuf, std::io::Error),
#[error("unable to read {0}: {1}")]
BlobRead(std::path::PathBuf, std::io::Error),
// TODO: proper error for blob finalize
#[error("unable to finalize blob {0}: {1}")]
BlobFinalize(std::path::PathBuf, std::io::Error),
}