mirror of
https://github.com/pesde-pkg/pesde.git
synced 2025-04-06 20:00:53 +01:00
feat: better install (#17)
Some checks are pending
Debug / Get build version (push) Waiting to run
Debug / Build for linux-x86_64 (push) Blocked by required conditions
Debug / Build for macos-aarch64 (push) Blocked by required conditions
Debug / Build for macos-x86_64 (push) Blocked by required conditions
Debug / Build for windows-x86_64 (push) Blocked by required conditions
Test & Lint / lint (push) Waiting to run
Some checks are pending
Debug / Get build version (push) Waiting to run
Debug / Build for linux-x86_64 (push) Blocked by required conditions
Debug / Build for macos-aarch64 (push) Blocked by required conditions
Debug / Build for macos-x86_64 (push) Blocked by required conditions
Debug / Build for windows-x86_64 (push) Blocked by required conditions
Test & Lint / lint (push) Waiting to run
* feat: better install * feat: support progress reporting for wally * chore: remove tracing-indicatif * chore: fix Cargo.toml * fix: indentation in bin link script * fix: spinner tick chars * feat: change progress message color * fix: remove pretty from fmt_layer Co-authored-by: dai <72147841+daimond113@users.noreply.github.com> * style: format code --------- Co-authored-by: dai <72147841+daimond113@users.noreply.github.com>
This commit is contained in:
parent
f009c957ca
commit
a41d9950f8
22 changed files with 1511 additions and 837 deletions
68
Cargo.lock
generated
68
Cargo.lock
generated
|
@ -357,12 +357,6 @@ version = "1.7.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "69f7f8c3906b62b754cd5326047894316021dcfe5a194c8ea52bdd94934a3457"
|
checksum = "69f7f8c3906b62b754cd5326047894316021dcfe5a194c8ea52bdd94934a3457"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "arrayvec"
|
|
||||||
version = "0.7.6"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "7c02d123df017efcdfbd739ef81735b36c5ba83ec3c59c80a9d7ecc718f92e50"
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "async-broadcast"
|
name = "async-broadcast"
|
||||||
version = "0.7.1"
|
version = "0.7.1"
|
||||||
|
@ -2880,7 +2874,6 @@ dependencies = [
|
||||||
"number_prefix",
|
"number_prefix",
|
||||||
"portable-atomic",
|
"portable-atomic",
|
||||||
"unicode-width 0.2.0",
|
"unicode-width 0.2.0",
|
||||||
"vt100",
|
|
||||||
"web-time",
|
"web-time",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
@ -3697,7 +3690,6 @@ dependencies = [
|
||||||
"toml",
|
"toml",
|
||||||
"toml_edit",
|
"toml_edit",
|
||||||
"tracing",
|
"tracing",
|
||||||
"tracing-indicatif",
|
|
||||||
"tracing-subscriber",
|
"tracing-subscriber",
|
||||||
"url",
|
"url",
|
||||||
"wax",
|
"wax",
|
||||||
|
@ -4147,10 +4139,12 @@ dependencies = [
|
||||||
"tokio",
|
"tokio",
|
||||||
"tokio-native-tls",
|
"tokio-native-tls",
|
||||||
"tokio-rustls",
|
"tokio-rustls",
|
||||||
|
"tokio-util",
|
||||||
"tower-service",
|
"tower-service",
|
||||||
"url",
|
"url",
|
||||||
"wasm-bindgen",
|
"wasm-bindgen",
|
||||||
"wasm-bindgen-futures",
|
"wasm-bindgen-futures",
|
||||||
|
"wasm-streams",
|
||||||
"web-sys",
|
"web-sys",
|
||||||
"webpki-roots",
|
"webpki-roots",
|
||||||
"windows-registry",
|
"windows-registry",
|
||||||
|
@ -5265,18 +5259,6 @@ dependencies = [
|
||||||
"valuable",
|
"valuable",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "tracing-indicatif"
|
|
||||||
version = "0.3.8"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "74ba258e9de86447f75edf6455fded8e5242704c6fccffe7bf8d7fb6daef1180"
|
|
||||||
dependencies = [
|
|
||||||
"indicatif",
|
|
||||||
"tracing",
|
|
||||||
"tracing-core",
|
|
||||||
"tracing-subscriber",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "tracing-log"
|
name = "tracing-log"
|
||||||
version = "0.2.0"
|
version = "0.2.0"
|
||||||
|
@ -5474,39 +5456,6 @@ version = "0.9.5"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "0b928f33d975fc6ad9f86c8f283853ad26bdd5b10b7f1542aa2fa15e2289105a"
|
checksum = "0b928f33d975fc6ad9f86c8f283853ad26bdd5b10b7f1542aa2fa15e2289105a"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "vt100"
|
|
||||||
version = "0.15.2"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "84cd863bf0db7e392ba3bd04994be3473491b31e66340672af5d11943c6274de"
|
|
||||||
dependencies = [
|
|
||||||
"itoa",
|
|
||||||
"log",
|
|
||||||
"unicode-width 0.1.14",
|
|
||||||
"vte",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "vte"
|
|
||||||
version = "0.11.1"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "f5022b5fbf9407086c180e9557be968742d839e68346af7792b8592489732197"
|
|
||||||
dependencies = [
|
|
||||||
"arrayvec",
|
|
||||||
"utf8parse",
|
|
||||||
"vte_generate_state_changes",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "vte_generate_state_changes"
|
|
||||||
version = "0.1.2"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "2e369bee1b05d510a7b4ed645f5faa90619e05437111783ea5848f28d97d3c2e"
|
|
||||||
dependencies = [
|
|
||||||
"proc-macro2",
|
|
||||||
"quote",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "walkdir"
|
name = "walkdir"
|
||||||
version = "2.5.0"
|
version = "2.5.0"
|
||||||
|
@ -5599,6 +5548,19 @@ version = "0.2.99"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "943aab3fdaaa029a6e0271b35ea10b72b943135afe9bffca82384098ad0e06a6"
|
checksum = "943aab3fdaaa029a6e0271b35ea10b72b943135afe9bffca82384098ad0e06a6"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "wasm-streams"
|
||||||
|
version = "0.4.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "15053d8d85c7eccdbefef60f06769760a563c7f0a9d6902a13d35c7800b0ad65"
|
||||||
|
dependencies = [
|
||||||
|
"futures-util",
|
||||||
|
"js-sys",
|
||||||
|
"wasm-bindgen",
|
||||||
|
"wasm-bindgen-futures",
|
||||||
|
"web-sys",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "wax"
|
name = "wax"
|
||||||
version = "0.6.0"
|
version = "0.6.0"
|
||||||
|
|
|
@ -16,7 +16,6 @@ bin = [
|
||||||
"dep:tracing-subscriber",
|
"dep:tracing-subscriber",
|
||||||
"reqwest/json",
|
"reqwest/json",
|
||||||
"dep:indicatif",
|
"dep:indicatif",
|
||||||
"dep:tracing-indicatif",
|
|
||||||
"dep:inquire",
|
"dep:inquire",
|
||||||
"dep:toml_edit",
|
"dep:toml_edit",
|
||||||
"dep:colored",
|
"dep:colored",
|
||||||
|
@ -49,7 +48,7 @@ toml = "0.8.19"
|
||||||
serde_with = "3.11.0"
|
serde_with = "3.11.0"
|
||||||
gix = { version = "0.68.0", default-features = false, features = ["blocking-http-transport-reqwest-rust-tls", "revparse-regex", "credentials", "parallel"] }
|
gix = { version = "0.68.0", default-features = false, features = ["blocking-http-transport-reqwest-rust-tls", "revparse-regex", "credentials", "parallel"] }
|
||||||
semver = { version = "1.0.24", features = ["serde"] }
|
semver = { version = "1.0.24", features = ["serde"] }
|
||||||
reqwest = { version = "0.12.9", default-features = false, features = ["rustls-tls"] }
|
reqwest = { version = "0.12.9", default-features = false, features = ["rustls-tls", "stream"] }
|
||||||
tokio-tar = "0.3.1"
|
tokio-tar = "0.3.1"
|
||||||
async-compression = { version = "0.4.18", features = ["tokio", "gzip"] }
|
async-compression = { version = "0.4.18", features = ["tokio", "gzip"] }
|
||||||
pathdiff = "0.2.3"
|
pathdiff = "0.2.3"
|
||||||
|
@ -83,7 +82,6 @@ clap = { version = "4.5.23", features = ["derive"], optional = true }
|
||||||
dirs = { version = "5.0.1", optional = true }
|
dirs = { version = "5.0.1", optional = true }
|
||||||
tracing-subscriber = { version = "0.3.19", features = ["env-filter"], optional = true }
|
tracing-subscriber = { version = "0.3.19", features = ["env-filter"], optional = true }
|
||||||
indicatif = { version = "0.17.9", optional = true }
|
indicatif = { version = "0.17.9", optional = true }
|
||||||
tracing-indicatif = { version = "0.3.8", optional = true }
|
|
||||||
inquire = { version = "0.7.5", optional = true }
|
inquire = { version = "0.7.5", optional = true }
|
||||||
|
|
||||||
[target.'cfg(target_os = "windows")'.dependencies]
|
[target.'cfg(target_os = "windows")'.dependencies]
|
||||||
|
|
|
@ -1,8 +1,15 @@
|
||||||
use crate::cli::{config::read_config, progress_bar, VersionedPackageName};
|
use crate::cli::{
|
||||||
|
config::read_config,
|
||||||
|
reporters::{self, CliReporter},
|
||||||
|
VersionedPackageName,
|
||||||
|
};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use clap::Args;
|
use clap::Args;
|
||||||
|
use colored::Colorize;
|
||||||
use fs_err::tokio as fs;
|
use fs_err::tokio as fs;
|
||||||
|
use indicatif::MultiProgress;
|
||||||
use pesde::{
|
use pesde::{
|
||||||
|
download_and_link::DownloadAndLinkOptions,
|
||||||
linking::generator::generate_bin_linking_module,
|
linking::generator::generate_bin_linking_module,
|
||||||
manifest::target::TargetKind,
|
manifest::target::TargetKind,
|
||||||
names::PackageName,
|
names::PackageName,
|
||||||
|
@ -14,7 +21,12 @@ use pesde::{
|
||||||
};
|
};
|
||||||
use semver::VersionReq;
|
use semver::VersionReq;
|
||||||
use std::{
|
use std::{
|
||||||
collections::HashSet, env::current_dir, ffi::OsString, io::Write, process::Command, sync::Arc,
|
collections::HashSet,
|
||||||
|
env::current_dir,
|
||||||
|
ffi::OsString,
|
||||||
|
io::{Stderr, Write},
|
||||||
|
process::Command,
|
||||||
|
sync::Arc,
|
||||||
};
|
};
|
||||||
use tokio::sync::Mutex;
|
use tokio::sync::Mutex;
|
||||||
|
|
||||||
|
@ -35,109 +47,123 @@ pub struct ExecuteCommand {
|
||||||
|
|
||||||
impl ExecuteCommand {
|
impl ExecuteCommand {
|
||||||
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
||||||
let index = match self.index {
|
let multi_progress = MultiProgress::new();
|
||||||
Some(index) => Some(index),
|
crate::PROGRESS_BARS
|
||||||
None => read_config().await.ok().map(|c| c.default_index),
|
.lock()
|
||||||
}
|
.unwrap()
|
||||||
.context("no index specified")?;
|
.replace(multi_progress.clone());
|
||||||
let source = PesdePackageSource::new(index);
|
|
||||||
source
|
|
||||||
.refresh(&project)
|
|
||||||
.await
|
|
||||||
.context("failed to refresh source")?;
|
|
||||||
|
|
||||||
let version_req = self.package.1.unwrap_or(VersionReq::STAR);
|
let (tempdir, bin_path) = reporters::run_with_reporter_and_writer(
|
||||||
let Some((version, pkg_ref)) = ('finder: {
|
std::io::stderr(),
|
||||||
let specifier = PesdeDependencySpecifier {
|
|multi_progress, root_progress, reporter| async {
|
||||||
name: self.package.0.clone(),
|
let multi_progress = multi_progress;
|
||||||
version: version_req.clone(),
|
let root_progress = root_progress;
|
||||||
index: None,
|
|
||||||
target: None,
|
|
||||||
};
|
|
||||||
|
|
||||||
if let Some(res) = source
|
root_progress.set_message("resolve");
|
||||||
.resolve(&specifier, &project, TargetKind::Lune, &mut HashSet::new())
|
|
||||||
.await
|
|
||||||
.context("failed to resolve package")?
|
|
||||||
.1
|
|
||||||
.pop_last()
|
|
||||||
{
|
|
||||||
break 'finder Some(res);
|
|
||||||
}
|
|
||||||
|
|
||||||
source
|
let index = match self.index {
|
||||||
.resolve(&specifier, &project, TargetKind::Luau, &mut HashSet::new())
|
Some(index) => Some(index),
|
||||||
.await
|
None => read_config().await.ok().map(|c| c.default_index),
|
||||||
.context("failed to resolve package")?
|
}
|
||||||
.1
|
.context("no index specified")?;
|
||||||
.pop_last()
|
let source = PesdePackageSource::new(index);
|
||||||
}) else {
|
source
|
||||||
anyhow::bail!(
|
.refresh(&project)
|
||||||
"no Lune or Luau package could be found for {}@{version_req}",
|
.await
|
||||||
self.package.0,
|
.context("failed to refresh source")?;
|
||||||
);
|
|
||||||
};
|
|
||||||
|
|
||||||
println!("using {}@{version}", pkg_ref.name);
|
let version_req = self.package.1.unwrap_or(VersionReq::STAR);
|
||||||
|
let Some((version, pkg_ref)) = ('finder: {
|
||||||
|
let specifier = PesdeDependencySpecifier {
|
||||||
|
name: self.package.0.clone(),
|
||||||
|
version: version_req.clone(),
|
||||||
|
index: None,
|
||||||
|
target: None,
|
||||||
|
};
|
||||||
|
|
||||||
let tmp_dir = project.cas_dir().join(".tmp");
|
if let Some(res) = source
|
||||||
fs::create_dir_all(&tmp_dir)
|
.resolve(&specifier, &project, TargetKind::Lune, &mut HashSet::new())
|
||||||
.await
|
.await
|
||||||
.context("failed to create temporary directory")?;
|
.context("failed to resolve package")?
|
||||||
let tempdir =
|
.1
|
||||||
tempfile::tempdir_in(tmp_dir).context("failed to create temporary directory")?;
|
.pop_last()
|
||||||
|
{
|
||||||
|
break 'finder Some(res);
|
||||||
|
}
|
||||||
|
|
||||||
let project = Project::new(
|
source
|
||||||
tempdir.path(),
|
.resolve(&specifier, &project, TargetKind::Luau, &mut HashSet::new())
|
||||||
None::<std::path::PathBuf>,
|
.await
|
||||||
project.data_dir(),
|
.context("failed to resolve package")?
|
||||||
project.cas_dir(),
|
.1
|
||||||
project.auth_config().clone(),
|
.pop_last()
|
||||||
);
|
}) else {
|
||||||
|
anyhow::bail!(
|
||||||
|
"no Lune or Luau package could be found for {}@{version_req}",
|
||||||
|
self.package.0,
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
let (fs, target) = source
|
let tmp_dir = project.cas_dir().join(".tmp");
|
||||||
.download(&pkg_ref, &project, &reqwest)
|
fs::create_dir_all(&tmp_dir)
|
||||||
.await
|
.await
|
||||||
.context("failed to download package")?;
|
.context("failed to create temporary directory")?;
|
||||||
let bin_path = target.bin_path().context("package has no binary export")?;
|
let tempdir = tempfile::tempdir_in(tmp_dir)
|
||||||
|
.context("failed to create temporary directory")?;
|
||||||
|
|
||||||
fs.write_to(tempdir.path(), project.cas_dir(), true)
|
let project = Project::new(
|
||||||
.await
|
tempdir.path(),
|
||||||
.context("failed to write package contents")?;
|
None::<std::path::PathBuf>,
|
||||||
|
project.data_dir(),
|
||||||
|
project.cas_dir(),
|
||||||
|
project.auth_config().clone(),
|
||||||
|
);
|
||||||
|
|
||||||
let mut refreshed_sources = HashSet::new();
|
let (fs, target) = source
|
||||||
|
.download(&pkg_ref, &project, &reqwest, Arc::new(()))
|
||||||
|
.await
|
||||||
|
.context("failed to download package")?;
|
||||||
|
let bin_path = target.bin_path().context("package has no binary export")?;
|
||||||
|
|
||||||
let graph = project
|
fs.write_to(tempdir.path(), project.cas_dir(), true)
|
||||||
.dependency_graph(None, &mut refreshed_sources, true)
|
.await
|
||||||
.await
|
.context("failed to write package contents")?;
|
||||||
.context("failed to build dependency graph")?;
|
|
||||||
let graph = Arc::new(graph);
|
|
||||||
|
|
||||||
let (rx, downloaded_graph) = project
|
let mut refreshed_sources = HashSet::new();
|
||||||
.download_and_link(
|
|
||||||
&graph,
|
|
||||||
&Arc::new(Mutex::new(refreshed_sources)),
|
|
||||||
&reqwest,
|
|
||||||
true,
|
|
||||||
true,
|
|
||||||
|_| async { Ok::<_, std::io::Error>(()) },
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.context("failed to download dependencies")?;
|
|
||||||
|
|
||||||
progress_bar(
|
let graph = project
|
||||||
graph.values().map(|versions| versions.len() as u64).sum(),
|
.dependency_graph(None, &mut refreshed_sources, true)
|
||||||
rx,
|
.await
|
||||||
"📥 ".to_string(),
|
.context("failed to build dependency graph")?;
|
||||||
"downloading dependencies".to_string(),
|
|
||||||
"downloaded dependencies".to_string(),
|
multi_progress.suspend(|| {
|
||||||
|
eprintln!(
|
||||||
|
"{}",
|
||||||
|
format!("using {}", format!("{}@{version}", pkg_ref.name).bold()).dimmed()
|
||||||
|
)
|
||||||
|
});
|
||||||
|
|
||||||
|
root_progress.reset();
|
||||||
|
root_progress.set_message("download");
|
||||||
|
root_progress.set_style(reporters::root_progress_style_with_progress());
|
||||||
|
|
||||||
|
project
|
||||||
|
.download_and_link(
|
||||||
|
&Arc::new(graph),
|
||||||
|
DownloadAndLinkOptions::<CliReporter<Stderr>, ()>::new(reqwest)
|
||||||
|
.reporter(reporter)
|
||||||
|
.refreshed_sources(Mutex::new(refreshed_sources))
|
||||||
|
.prod(true)
|
||||||
|
.write(true),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.context("failed to download and link dependencies")?;
|
||||||
|
|
||||||
|
anyhow::Ok((tempdir, bin_path.clone()))
|
||||||
|
},
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
|
|
||||||
downloaded_graph
|
|
||||||
.await
|
|
||||||
.context("failed to download & link dependencies")?;
|
|
||||||
|
|
||||||
let mut caller =
|
let mut caller =
|
||||||
tempfile::NamedTempFile::new_in(tempdir.path()).context("failed to create tempfile")?;
|
tempfile::NamedTempFile::new_in(tempdir.path()).context("failed to create tempfile")?;
|
||||||
caller
|
caller
|
||||||
|
|
|
@ -1,20 +1,10 @@
|
||||||
use crate::cli::{
|
use crate::cli::{
|
||||||
bin_dir, files::make_executable, progress_bar, run_on_workspace_members, up_to_date_lockfile,
|
install::{install, InstallOptions},
|
||||||
|
run_on_workspace_members,
|
||||||
};
|
};
|
||||||
use anyhow::Context;
|
|
||||||
use clap::Args;
|
use clap::Args;
|
||||||
use colored::{ColoredString, Colorize};
|
use pesde::Project;
|
||||||
use fs_err::tokio as fs;
|
use std::num::NonZeroUsize;
|
||||||
use futures::future::try_join_all;
|
|
||||||
use pesde::{
|
|
||||||
download_and_link::filter_graph, lockfile::Lockfile, manifest::target::TargetKind, Project,
|
|
||||||
MANIFEST_FILE_NAME,
|
|
||||||
};
|
|
||||||
use std::{
|
|
||||||
collections::{BTreeSet, HashMap, HashSet},
|
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
use tokio::sync::Mutex;
|
|
||||||
|
|
||||||
#[derive(Debug, Args, Copy, Clone)]
|
#[derive(Debug, Args, Copy, Clone)]
|
||||||
pub struct InstallCommand {
|
pub struct InstallCommand {
|
||||||
|
@ -25,303 +15,35 @@ pub struct InstallCommand {
|
||||||
/// Whether to not install dev dependencies
|
/// Whether to not install dev dependencies
|
||||||
#[arg(long)]
|
#[arg(long)]
|
||||||
prod: bool,
|
prod: bool,
|
||||||
}
|
|
||||||
|
|
||||||
fn bin_link_file(alias: &str) -> String {
|
/// The maximum number of concurrent network requests
|
||||||
let mut all_combinations = BTreeSet::new();
|
#[arg(long, default_value = "16")]
|
||||||
|
network_concurrency: NonZeroUsize,
|
||||||
for a in TargetKind::VARIANTS {
|
|
||||||
for b in TargetKind::VARIANTS {
|
|
||||||
all_combinations.insert((a, b));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let all_folders = all_combinations
|
|
||||||
.into_iter()
|
|
||||||
.map(|(a, b)| format!("{:?}", a.packages_folder(b)))
|
|
||||||
.collect::<BTreeSet<_>>()
|
|
||||||
.into_iter()
|
|
||||||
.collect::<Vec<_>>()
|
|
||||||
.join(", ");
|
|
||||||
|
|
||||||
format!(
|
|
||||||
r#"local process = require("@lune/process")
|
|
||||||
local fs = require("@lune/fs")
|
|
||||||
local stdio = require("@lune/stdio")
|
|
||||||
|
|
||||||
local project_root = process.cwd
|
|
||||||
local path_components = string.split(string.gsub(project_root, "\\", "/"), "/")
|
|
||||||
|
|
||||||
for i = #path_components, 1, -1 do
|
|
||||||
local path = table.concat(path_components, "/", 1, i)
|
|
||||||
if fs.isFile(path .. "/{MANIFEST_FILE_NAME}") then
|
|
||||||
project_root = path
|
|
||||||
break
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
for _, packages_folder in {{ {all_folders} }} do
|
|
||||||
local path = `{{project_root}}/{{packages_folder}}/{alias}.bin.luau`
|
|
||||||
|
|
||||||
if fs.isFile(path) then
|
|
||||||
require(path)
|
|
||||||
return
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
stdio.ewrite(stdio.color("red") .. "binary `{alias}` not found. are you in the right directory?" .. stdio.color("reset") .. "\n")
|
|
||||||
"#,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(feature = "patches")]
|
|
||||||
const JOBS: u8 = 5;
|
|
||||||
#[cfg(not(feature = "patches"))]
|
|
||||||
const JOBS: u8 = 4;
|
|
||||||
|
|
||||||
fn job(n: u8) -> ColoredString {
|
|
||||||
format!("[{n}/{JOBS}]").dimmed().bold()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, thiserror::Error)]
|
#[derive(Debug, thiserror::Error)]
|
||||||
#[error(transparent)]
|
#[error(transparent)]
|
||||||
struct CallbackError(#[from] anyhow::Error);
|
struct CallbackError(#[from] anyhow::Error);
|
||||||
|
|
||||||
impl InstallCommand {
|
impl InstallCommand {
|
||||||
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
||||||
let mut refreshed_sources = HashSet::new();
|
let options = InstallOptions {
|
||||||
|
locked: self.locked,
|
||||||
let manifest = project
|
prod: self.prod,
|
||||||
.deser_manifest()
|
write: true,
|
||||||
.await
|
network_concurrency: self.network_concurrency,
|
||||||
.context("failed to read manifest")?;
|
use_lockfile: true,
|
||||||
|
|
||||||
let lockfile = if self.locked {
|
|
||||||
match up_to_date_lockfile(&project).await? {
|
|
||||||
None => {
|
|
||||||
anyhow::bail!(
|
|
||||||
"lockfile is out of sync, run `{} install` to update it",
|
|
||||||
env!("CARGO_BIN_NAME")
|
|
||||||
);
|
|
||||||
}
|
|
||||||
file => file,
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
match project.deser_lockfile().await {
|
|
||||||
Ok(lockfile) => {
|
|
||||||
if lockfile.overrides != manifest.overrides {
|
|
||||||
tracing::debug!("overrides are different");
|
|
||||||
None
|
|
||||||
} else if lockfile.target != manifest.target.kind() {
|
|
||||||
tracing::debug!("target kind is different");
|
|
||||||
None
|
|
||||||
} else {
|
|
||||||
Some(lockfile)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(pesde::errors::LockfileReadError::Io(e))
|
|
||||||
if e.kind() == std::io::ErrorKind::NotFound =>
|
|
||||||
{
|
|
||||||
None
|
|
||||||
}
|
|
||||||
Err(e) => return Err(e.into()),
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
|
|
||||||
println!(
|
install(&options, &project, reqwest.clone(), true).await?;
|
||||||
"\n{}\n",
|
|
||||||
format!("[now installing {} {}]", manifest.name, manifest.target)
|
|
||||||
.bold()
|
|
||||||
.on_bright_black()
|
|
||||||
);
|
|
||||||
|
|
||||||
println!("{} ❌ removing current package folders", job(1));
|
run_on_workspace_members(&project, |project| {
|
||||||
|
let reqwest = reqwest.clone();
|
||||||
{
|
async move {
|
||||||
let mut deleted_folders = HashMap::new();
|
install(&options, &project, reqwest, false).await?;
|
||||||
|
Ok(())
|
||||||
for target_kind in TargetKind::VARIANTS {
|
|
||||||
let folder = manifest.target.kind().packages_folder(target_kind);
|
|
||||||
let package_dir = project.package_dir();
|
|
||||||
|
|
||||||
deleted_folders
|
|
||||||
.entry(folder.to_string())
|
|
||||||
.or_insert_with(|| async move {
|
|
||||||
tracing::debug!("deleting the {folder} folder");
|
|
||||||
|
|
||||||
if let Some(e) = fs::remove_dir_all(package_dir.join(&folder))
|
|
||||||
.await
|
|
||||||
.err()
|
|
||||||
.filter(|e| e.kind() != std::io::ErrorKind::NotFound)
|
|
||||||
{
|
|
||||||
return Err(e).context(format!("failed to remove the {folder} folder"));
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
|
})
|
||||||
try_join_all(deleted_folders.into_values())
|
.await?;
|
||||||
.await
|
|
||||||
.context("failed to remove package folders")?;
|
|
||||||
}
|
|
||||||
|
|
||||||
let old_graph = lockfile.map(|lockfile| {
|
|
||||||
lockfile
|
|
||||||
.graph
|
|
||||||
.into_iter()
|
|
||||||
.map(|(name, versions)| {
|
|
||||||
(
|
|
||||||
name,
|
|
||||||
versions
|
|
||||||
.into_iter()
|
|
||||||
.map(|(version, node)| (version, node.node))
|
|
||||||
.collect(),
|
|
||||||
)
|
|
||||||
})
|
|
||||||
.collect()
|
|
||||||
});
|
|
||||||
|
|
||||||
println!("{} 📦 building dependency graph", job(2));
|
|
||||||
|
|
||||||
let graph = project
|
|
||||||
.dependency_graph(old_graph.as_ref(), &mut refreshed_sources, false)
|
|
||||||
.await
|
|
||||||
.context("failed to build dependency graph")?;
|
|
||||||
let graph = Arc::new(graph);
|
|
||||||
|
|
||||||
let bin_folder = bin_dir().await?;
|
|
||||||
|
|
||||||
let downloaded_graph = {
|
|
||||||
let (rx, downloaded_graph) = project
|
|
||||||
.download_and_link(
|
|
||||||
&graph,
|
|
||||||
&Arc::new(Mutex::new(refreshed_sources)),
|
|
||||||
&reqwest,
|
|
||||||
self.prod,
|
|
||||||
true,
|
|
||||||
|graph| {
|
|
||||||
let graph = graph.clone();
|
|
||||||
|
|
||||||
async move {
|
|
||||||
try_join_all(
|
|
||||||
graph
|
|
||||||
.values()
|
|
||||||
.flat_map(|versions| versions.values())
|
|
||||||
.filter(|node| node.target.bin_path().is_some())
|
|
||||||
.filter_map(|node| node.node.direct.as_ref())
|
|
||||||
.map(|(alias, _, _)| alias)
|
|
||||||
.filter(|alias| {
|
|
||||||
if *alias == env!("CARGO_BIN_NAME") {
|
|
||||||
tracing::warn!(
|
|
||||||
"package {alias} has the same name as the CLI, skipping bin link"
|
|
||||||
);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
true
|
|
||||||
})
|
|
||||||
.map(|alias| {
|
|
||||||
let bin_folder = bin_folder.clone();
|
|
||||||
async move {
|
|
||||||
let bin_exec_file = bin_folder.join(alias).with_extension(std::env::consts::EXE_EXTENSION);
|
|
||||||
|
|
||||||
let impl_folder = bin_folder.join(".impl");
|
|
||||||
fs::create_dir_all(&impl_folder).await.context("failed to create bin link folder")?;
|
|
||||||
|
|
||||||
let bin_file = impl_folder.join(alias).with_extension("luau");
|
|
||||||
fs::write(&bin_file, bin_link_file(alias))
|
|
||||||
.await
|
|
||||||
.context("failed to write bin link file")?;
|
|
||||||
|
|
||||||
|
|
||||||
#[cfg(windows)]
|
|
||||||
{
|
|
||||||
fs::copy(
|
|
||||||
std::env::current_exe()
|
|
||||||
.context("failed to get current executable path")?,
|
|
||||||
&bin_exec_file,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.context("failed to copy bin link file")?;
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(not(windows))]
|
|
||||||
{
|
|
||||||
fs::write(
|
|
||||||
&bin_exec_file,
|
|
||||||
format!(r#"#!/bin/sh
|
|
||||||
exec lune run "$(dirname "$0")/.impl/{alias}.luau" -- "$@""#
|
|
||||||
),
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.context("failed to link bin link file")?;
|
|
||||||
}
|
|
||||||
|
|
||||||
make_executable(&bin_exec_file).await.context("failed to make bin link file executable")?;
|
|
||||||
|
|
||||||
Ok::<_, CallbackError>(())
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.map(|_| ())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.context("failed to download dependencies")?;
|
|
||||||
|
|
||||||
progress_bar(
|
|
||||||
graph.values().map(|versions| versions.len() as u64).sum(),
|
|
||||||
rx,
|
|
||||||
format!("{} 📥 ", job(3)),
|
|
||||||
"downloading dependencies".to_string(),
|
|
||||||
"downloaded dependencies".to_string(),
|
|
||||||
)
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
downloaded_graph
|
|
||||||
.await
|
|
||||||
.context("failed to download & link dependencies")?
|
|
||||||
};
|
|
||||||
|
|
||||||
#[cfg(feature = "patches")]
|
|
||||||
{
|
|
||||||
let rx = project
|
|
||||||
.apply_patches(&filter_graph(&downloaded_graph, self.prod))
|
|
||||||
.await
|
|
||||||
.context("failed to apply patches")?;
|
|
||||||
|
|
||||||
progress_bar(
|
|
||||||
manifest.patches.values().map(|v| v.len() as u64).sum(),
|
|
||||||
rx,
|
|
||||||
format!("{} 🩹 ", job(JOBS - 1)),
|
|
||||||
"applying patches".to_string(),
|
|
||||||
"applied patches".to_string(),
|
|
||||||
)
|
|
||||||
.await?;
|
|
||||||
}
|
|
||||||
|
|
||||||
println!("{} 🧹 finishing up", job(JOBS));
|
|
||||||
|
|
||||||
project
|
|
||||||
.write_lockfile(Lockfile {
|
|
||||||
name: manifest.name,
|
|
||||||
version: manifest.version,
|
|
||||||
target: manifest.target.kind(),
|
|
||||||
overrides: manifest.overrides,
|
|
||||||
|
|
||||||
graph: downloaded_graph,
|
|
||||||
|
|
||||||
workspace: run_on_workspace_members(&project, |project| {
|
|
||||||
let reqwest = reqwest.clone();
|
|
||||||
async move { Box::pin(self.run(project, reqwest)).await }
|
|
||||||
})
|
|
||||||
.await?,
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.context("failed to write lockfile")?;
|
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,3 +1,5 @@
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use crate::cli::{up_to_date_lockfile, VersionedPackageName};
|
use crate::cli::{up_to_date_lockfile, VersionedPackageName};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use clap::Args;
|
use clap::Args;
|
||||||
|
@ -49,7 +51,7 @@ impl PatchCommand {
|
||||||
fs::create_dir_all(&directory).await?;
|
fs::create_dir_all(&directory).await?;
|
||||||
|
|
||||||
source
|
source
|
||||||
.download(&node.node.pkg_ref, &project, &reqwest)
|
.download(&node.node.pkg_ref, &project, &reqwest, Arc::new(()))
|
||||||
.await?
|
.await?
|
||||||
.0
|
.0
|
||||||
.write_to(&directory, project.cas_dir(), false)
|
.write_to(&directory, project.cas_dir(), false)
|
||||||
|
|
|
@ -1,84 +1,42 @@
|
||||||
use crate::cli::{progress_bar, run_on_workspace_members};
|
use crate::cli::{
|
||||||
use anyhow::Context;
|
install::{install, InstallOptions},
|
||||||
|
run_on_workspace_members,
|
||||||
|
};
|
||||||
use clap::Args;
|
use clap::Args;
|
||||||
use colored::Colorize;
|
use pesde::Project;
|
||||||
use pesde::{lockfile::Lockfile, Project};
|
use std::num::NonZeroUsize;
|
||||||
use std::{collections::HashSet, sync::Arc};
|
|
||||||
use tokio::sync::Mutex;
|
|
||||||
|
|
||||||
#[derive(Debug, Args, Copy, Clone)]
|
#[derive(Debug, Args, Copy, Clone)]
|
||||||
pub struct UpdateCommand {}
|
pub struct UpdateCommand {
|
||||||
|
/// Update the dependencies but don't install them
|
||||||
|
#[arg(long)]
|
||||||
|
no_install: bool,
|
||||||
|
|
||||||
|
/// The maximum number of concurrent network requests
|
||||||
|
#[arg(long, default_value = "16")]
|
||||||
|
network_concurrency: NonZeroUsize,
|
||||||
|
}
|
||||||
|
|
||||||
impl UpdateCommand {
|
impl UpdateCommand {
|
||||||
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
pub async fn run(self, project: Project, reqwest: reqwest::Client) -> anyhow::Result<()> {
|
||||||
let mut refreshed_sources = HashSet::new();
|
let options = InstallOptions {
|
||||||
|
locked: false,
|
||||||
|
prod: false,
|
||||||
|
write: !self.no_install,
|
||||||
|
network_concurrency: self.network_concurrency,
|
||||||
|
use_lockfile: false,
|
||||||
|
};
|
||||||
|
|
||||||
let manifest = project
|
install(&options, &project, reqwest.clone(), true).await?;
|
||||||
.deser_manifest()
|
|
||||||
.await
|
|
||||||
.context("failed to read manifest")?;
|
|
||||||
|
|
||||||
println!(
|
run_on_workspace_members(&project, |project| {
|
||||||
"\n{}\n",
|
let reqwest = reqwest.clone();
|
||||||
format!("[now updating {} {}]", manifest.name, manifest.target)
|
async move {
|
||||||
.bold()
|
install(&options, &project, reqwest, false).await?;
|
||||||
.on_bright_black()
|
Ok(())
|
||||||
);
|
}
|
||||||
|
})
|
||||||
let graph = project
|
.await?;
|
||||||
.dependency_graph(None, &mut refreshed_sources, false)
|
|
||||||
.await
|
|
||||||
.context("failed to build dependency graph")?;
|
|
||||||
let graph = Arc::new(graph);
|
|
||||||
|
|
||||||
project
|
|
||||||
.write_lockfile(Lockfile {
|
|
||||||
name: manifest.name,
|
|
||||||
version: manifest.version,
|
|
||||||
target: manifest.target.kind(),
|
|
||||||
overrides: manifest.overrides,
|
|
||||||
|
|
||||||
graph: {
|
|
||||||
let (rx, downloaded_graph) = project
|
|
||||||
.download_and_link(
|
|
||||||
&graph,
|
|
||||||
&Arc::new(Mutex::new(refreshed_sources)),
|
|
||||||
&reqwest,
|
|
||||||
false,
|
|
||||||
false,
|
|
||||||
|_| async { Ok::<_, std::io::Error>(()) },
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
.context("failed to download dependencies")?;
|
|
||||||
|
|
||||||
progress_bar(
|
|
||||||
graph.values().map(|versions| versions.len() as u64).sum(),
|
|
||||||
rx,
|
|
||||||
"📥 ".to_string(),
|
|
||||||
"downloading dependencies".to_string(),
|
|
||||||
"downloaded dependencies".to_string(),
|
|
||||||
)
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
downloaded_graph
|
|
||||||
.await
|
|
||||||
.context("failed to download dependencies")?
|
|
||||||
},
|
|
||||||
|
|
||||||
workspace: run_on_workspace_members(&project, |project| {
|
|
||||||
let reqwest = reqwest.clone();
|
|
||||||
async move { Box::pin(self.run(project, reqwest)).await }
|
|
||||||
})
|
|
||||||
.await?,
|
|
||||||
})
|
|
||||||
.await
|
|
||||||
.context("failed to write lockfile")?;
|
|
||||||
|
|
||||||
println!(
|
|
||||||
"\n\n{}. run `{} install` in order to install the new dependencies",
|
|
||||||
"✅ done".green(),
|
|
||||||
env!("CARGO_BIN_NAME")
|
|
||||||
);
|
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
482
src/cli/install.rs
Normal file
482
src/cli/install.rs
Normal file
|
@ -0,0 +1,482 @@
|
||||||
|
use std::{
|
||||||
|
collections::{BTreeMap, BTreeSet, HashMap, HashSet},
|
||||||
|
num::NonZeroUsize,
|
||||||
|
sync::Arc,
|
||||||
|
time::Instant,
|
||||||
|
};
|
||||||
|
|
||||||
|
use anyhow::Context;
|
||||||
|
use colored::Colorize;
|
||||||
|
use fs_err::tokio as fs;
|
||||||
|
use futures::future::try_join_all;
|
||||||
|
use pesde::{
|
||||||
|
download_and_link::{filter_graph, DownloadAndLinkHooks, DownloadAndLinkOptions},
|
||||||
|
lockfile::{DependencyGraph, DownloadedGraph, Lockfile},
|
||||||
|
manifest::{target::TargetKind, DependencyType},
|
||||||
|
Project, MANIFEST_FILE_NAME,
|
||||||
|
};
|
||||||
|
use tokio::{sync::Mutex, task::JoinSet};
|
||||||
|
|
||||||
|
use crate::cli::{
|
||||||
|
bin_dir,
|
||||||
|
reporters::{self, CliReporter},
|
||||||
|
run_on_workspace_members, up_to_date_lockfile,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::files::make_executable;
|
||||||
|
|
||||||
|
fn bin_link_file(alias: &str) -> String {
|
||||||
|
let mut all_combinations = BTreeSet::new();
|
||||||
|
|
||||||
|
for a in TargetKind::VARIANTS {
|
||||||
|
for b in TargetKind::VARIANTS {
|
||||||
|
all_combinations.insert((a, b));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let all_folders = all_combinations
|
||||||
|
.into_iter()
|
||||||
|
.map(|(a, b)| format!("{:?}", a.packages_folder(b)))
|
||||||
|
.collect::<BTreeSet<_>>()
|
||||||
|
.into_iter()
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
.join(", ");
|
||||||
|
|
||||||
|
format!(
|
||||||
|
r#"local process = require("@lune/process")
|
||||||
|
local fs = require("@lune/fs")
|
||||||
|
local stdio = require("@lune/stdio")
|
||||||
|
|
||||||
|
local project_root = process.cwd
|
||||||
|
local path_components = string.split(string.gsub(project_root, "\\", "/"), "/")
|
||||||
|
|
||||||
|
for i = #path_components, 1, -1 do
|
||||||
|
local path = table.concat(path_components, "/", 1, i)
|
||||||
|
if fs.isFile(path .. "/{MANIFEST_FILE_NAME}") then
|
||||||
|
project_root = path
|
||||||
|
break
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
for _, packages_folder in {{ {all_folders} }} do
|
||||||
|
local path = `{{project_root}}/{{packages_folder}}/{alias}.bin.luau`
|
||||||
|
|
||||||
|
if fs.isFile(path) then
|
||||||
|
require(path)
|
||||||
|
return
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
stdio.ewrite(stdio.color("red") .. "binary `{alias}` not found. are you in the right directory?" .. stdio.color("reset") .. "\n")
|
||||||
|
"#,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct InstallHooks {
|
||||||
|
pub bin_folder: std::path::PathBuf,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, thiserror::Error)]
|
||||||
|
#[error(transparent)]
|
||||||
|
pub struct InstallHooksError(#[from] anyhow::Error);
|
||||||
|
|
||||||
|
impl DownloadAndLinkHooks for InstallHooks {
|
||||||
|
type Error = InstallHooksError;
|
||||||
|
|
||||||
|
async fn on_bins_downloaded(
|
||||||
|
&self,
|
||||||
|
downloaded_graph: &pesde::lockfile::DownloadedGraph,
|
||||||
|
) -> Result<(), Self::Error> {
|
||||||
|
let mut tasks = downloaded_graph
|
||||||
|
.values()
|
||||||
|
.flat_map(|versions| versions.values())
|
||||||
|
.filter(|node| node.target.bin_path().is_some())
|
||||||
|
.filter_map(|node| node.node.direct.as_ref())
|
||||||
|
.map(|(alias, _, _)| alias)
|
||||||
|
.filter(|alias| {
|
||||||
|
if *alias == env!("CARGO_BIN_NAME") {
|
||||||
|
tracing::warn!(
|
||||||
|
"package {alias} has the same name as the CLI, skipping bin link"
|
||||||
|
);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
true
|
||||||
|
})
|
||||||
|
.map(|alias| {
|
||||||
|
let bin_folder = self.bin_folder.clone();
|
||||||
|
let alias = alias.clone();
|
||||||
|
|
||||||
|
async move {
|
||||||
|
let bin_exec_file = bin_folder
|
||||||
|
.join(&alias)
|
||||||
|
.with_extension(std::env::consts::EXE_EXTENSION);
|
||||||
|
|
||||||
|
let impl_folder = bin_folder.join(".impl");
|
||||||
|
fs::create_dir_all(&impl_folder)
|
||||||
|
.await
|
||||||
|
.context("failed to create bin link folder")?;
|
||||||
|
|
||||||
|
let bin_file = impl_folder.join(&alias).with_extension("luau");
|
||||||
|
fs::write(&bin_file, bin_link_file(&alias))
|
||||||
|
.await
|
||||||
|
.context("failed to write bin link file")?;
|
||||||
|
|
||||||
|
#[cfg(windows)]
|
||||||
|
match fs::symlink_file(
|
||||||
|
std::env::current_exe().context("failed to get current executable path")?,
|
||||||
|
&bin_exec_file,
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(_) => {}
|
||||||
|
Err(e) if e.kind() == std::io::ErrorKind::AlreadyExists => {}
|
||||||
|
e => e.context("failed to copy bin link file")?,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(not(windows))]
|
||||||
|
fs::write(
|
||||||
|
&bin_exec_file,
|
||||||
|
format!(
|
||||||
|
r#"#!/bin/sh
|
||||||
|
exec lune run "$(dirname "$0")/.impl/{alias}.luau" -- "$@""#
|
||||||
|
),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.context("failed to link bin link file")?;
|
||||||
|
|
||||||
|
make_executable(&bin_exec_file)
|
||||||
|
.await
|
||||||
|
.context("failed to make bin link file executable")?;
|
||||||
|
|
||||||
|
Ok::<_, anyhow::Error>(())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.collect::<JoinSet<_>>();
|
||||||
|
|
||||||
|
while let Some(task) = tasks.join_next().await {
|
||||||
|
task.unwrap()?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Copy)]
|
||||||
|
pub struct InstallOptions {
|
||||||
|
pub locked: bool,
|
||||||
|
pub prod: bool,
|
||||||
|
pub write: bool,
|
||||||
|
pub use_lockfile: bool,
|
||||||
|
pub network_concurrency: NonZeroUsize,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn install(
|
||||||
|
options: &InstallOptions,
|
||||||
|
project: &Project,
|
||||||
|
reqwest: reqwest::Client,
|
||||||
|
is_root: bool,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
let start = Instant::now();
|
||||||
|
|
||||||
|
let mut refreshed_sources = HashSet::new();
|
||||||
|
|
||||||
|
let manifest = project
|
||||||
|
.deser_manifest()
|
||||||
|
.await
|
||||||
|
.context("failed to read manifest")?;
|
||||||
|
|
||||||
|
let lockfile = if options.locked {
|
||||||
|
match up_to_date_lockfile(project).await? {
|
||||||
|
None => {
|
||||||
|
anyhow::bail!(
|
||||||
|
"lockfile is out of sync, run `{} install` to update it",
|
||||||
|
env!("CARGO_BIN_NAME")
|
||||||
|
);
|
||||||
|
}
|
||||||
|
file => file,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
match project.deser_lockfile().await {
|
||||||
|
Ok(lockfile) => {
|
||||||
|
if lockfile.overrides != manifest.overrides {
|
||||||
|
tracing::debug!("overrides are different");
|
||||||
|
None
|
||||||
|
} else if lockfile.target != manifest.target.kind() {
|
||||||
|
tracing::debug!("target kind is different");
|
||||||
|
None
|
||||||
|
} else {
|
||||||
|
Some(lockfile)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(pesde::errors::LockfileReadError::Io(e))
|
||||||
|
if e.kind() == std::io::ErrorKind::NotFound =>
|
||||||
|
{
|
||||||
|
None
|
||||||
|
}
|
||||||
|
Err(e) => return Err(e.into()),
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let (new_lockfile, old_graph) =
|
||||||
|
reporters::run_with_reporter(|_, root_progress, reporter| async {
|
||||||
|
let root_progress = root_progress;
|
||||||
|
|
||||||
|
root_progress.set_prefix(format!("{} {}: ", manifest.name, manifest.target));
|
||||||
|
root_progress.set_message("clean");
|
||||||
|
|
||||||
|
if options.write {
|
||||||
|
let mut deleted_folders = HashMap::new();
|
||||||
|
|
||||||
|
for target_kind in TargetKind::VARIANTS {
|
||||||
|
let folder = manifest.target.kind().packages_folder(target_kind);
|
||||||
|
let package_dir = project.package_dir();
|
||||||
|
|
||||||
|
deleted_folders
|
||||||
|
.entry(folder.to_string())
|
||||||
|
.or_insert_with(|| async move {
|
||||||
|
tracing::debug!("deleting the {folder} folder");
|
||||||
|
|
||||||
|
if let Some(e) = fs::remove_dir_all(package_dir.join(&folder))
|
||||||
|
.await
|
||||||
|
.err()
|
||||||
|
.filter(|e| e.kind() != std::io::ErrorKind::NotFound)
|
||||||
|
{
|
||||||
|
return Err(e)
|
||||||
|
.context(format!("failed to remove the {folder} folder"));
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
try_join_all(deleted_folders.into_values())
|
||||||
|
.await
|
||||||
|
.context("failed to remove package folders")?;
|
||||||
|
}
|
||||||
|
|
||||||
|
root_progress.reset();
|
||||||
|
root_progress.set_message("resolve");
|
||||||
|
|
||||||
|
let old_graph = lockfile.map(|lockfile| {
|
||||||
|
lockfile
|
||||||
|
.graph
|
||||||
|
.into_iter()
|
||||||
|
.map(|(name, versions)| {
|
||||||
|
(
|
||||||
|
name,
|
||||||
|
versions
|
||||||
|
.into_iter()
|
||||||
|
.map(|(version, node)| (version, node.node))
|
||||||
|
.collect(),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
.collect()
|
||||||
|
});
|
||||||
|
|
||||||
|
let graph = project
|
||||||
|
.dependency_graph(
|
||||||
|
old_graph.as_ref().filter(|_| options.use_lockfile),
|
||||||
|
&mut refreshed_sources,
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.context("failed to build dependency graph")?;
|
||||||
|
let graph = Arc::new(graph);
|
||||||
|
|
||||||
|
root_progress.reset();
|
||||||
|
root_progress.set_length(0);
|
||||||
|
root_progress.set_message("download");
|
||||||
|
root_progress.set_style(reporters::root_progress_style_with_progress());
|
||||||
|
|
||||||
|
let hooks = InstallHooks {
|
||||||
|
bin_folder: bin_dir().await?,
|
||||||
|
};
|
||||||
|
|
||||||
|
let downloaded_graph = project
|
||||||
|
.download_and_link(
|
||||||
|
&graph,
|
||||||
|
DownloadAndLinkOptions::<CliReporter, InstallHooks>::new(reqwest.clone())
|
||||||
|
.reporter(reporter.clone())
|
||||||
|
.hooks(hooks)
|
||||||
|
.refreshed_sources(Mutex::new(refreshed_sources))
|
||||||
|
.prod(options.prod)
|
||||||
|
.write(options.write)
|
||||||
|
.network_concurrency(options.network_concurrency),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.context("failed to download and link dependencies")?;
|
||||||
|
|
||||||
|
#[cfg(feature = "patches")]
|
||||||
|
if options.write {
|
||||||
|
root_progress.reset();
|
||||||
|
root_progress.set_length(0);
|
||||||
|
root_progress.set_message("patch");
|
||||||
|
|
||||||
|
project
|
||||||
|
.apply_patches(&filter_graph(&downloaded_graph, options.prod), reporter)
|
||||||
|
.await?;
|
||||||
|
}
|
||||||
|
|
||||||
|
root_progress.set_message("finish");
|
||||||
|
|
||||||
|
let new_lockfile = Lockfile {
|
||||||
|
name: manifest.name.clone(),
|
||||||
|
version: manifest.version,
|
||||||
|
target: manifest.target.kind(),
|
||||||
|
overrides: manifest.overrides,
|
||||||
|
|
||||||
|
graph: downloaded_graph,
|
||||||
|
|
||||||
|
workspace: run_on_workspace_members(project, |_| async { Ok(()) }).await?,
|
||||||
|
};
|
||||||
|
|
||||||
|
project
|
||||||
|
.write_lockfile(&new_lockfile)
|
||||||
|
.await
|
||||||
|
.context("failed to write lockfile")?;
|
||||||
|
|
||||||
|
anyhow::Ok((new_lockfile, old_graph.unwrap_or_default()))
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
let elapsed = start.elapsed();
|
||||||
|
|
||||||
|
if is_root {
|
||||||
|
println!();
|
||||||
|
}
|
||||||
|
|
||||||
|
print_package_diff(
|
||||||
|
&format!("{} {}:", manifest.name, manifest.target),
|
||||||
|
old_graph,
|
||||||
|
new_lockfile.graph,
|
||||||
|
);
|
||||||
|
|
||||||
|
println!("done in {:.2}s", elapsed.as_secs_f64());
|
||||||
|
println!();
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prints the difference between two graphs.
|
||||||
|
pub fn print_package_diff(prefix: &str, old_graph: DependencyGraph, new_graph: DownloadedGraph) {
|
||||||
|
let mut old_pkg_map = BTreeMap::new();
|
||||||
|
let mut old_direct_pkg_map = BTreeMap::new();
|
||||||
|
let mut new_pkg_map = BTreeMap::new();
|
||||||
|
let mut new_direct_pkg_map = BTreeMap::new();
|
||||||
|
|
||||||
|
for (name, versions) in &old_graph {
|
||||||
|
for (version, node) in versions {
|
||||||
|
old_pkg_map.insert((name.clone(), version), node);
|
||||||
|
if node.direct.is_some() {
|
||||||
|
old_direct_pkg_map.insert((name.clone(), version), node);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (name, versions) in &new_graph {
|
||||||
|
for (version, node) in versions {
|
||||||
|
new_pkg_map.insert((name.clone(), version), &node.node);
|
||||||
|
if node.node.direct.is_some() {
|
||||||
|
new_direct_pkg_map.insert((name.clone(), version), &node.node);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let added_pkgs = new_pkg_map
|
||||||
|
.iter()
|
||||||
|
.filter(|(key, _)| !old_pkg_map.contains_key(key))
|
||||||
|
.map(|(key, &node)| (key, node))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let removed_pkgs = old_pkg_map
|
||||||
|
.iter()
|
||||||
|
.filter(|(key, _)| !new_pkg_map.contains_key(key))
|
||||||
|
.map(|(key, &node)| (key, node))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let added_direct_pkgs = new_direct_pkg_map
|
||||||
|
.iter()
|
||||||
|
.filter(|(key, _)| !old_direct_pkg_map.contains_key(key))
|
||||||
|
.map(|(key, &node)| (key, node))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let removed_direct_pkgs = old_direct_pkg_map
|
||||||
|
.iter()
|
||||||
|
.filter(|(key, _)| !new_direct_pkg_map.contains_key(key))
|
||||||
|
.map(|(key, &node)| (key, node))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
|
let prefix = prefix.bold();
|
||||||
|
|
||||||
|
let no_changes = added_pkgs.is_empty()
|
||||||
|
&& removed_pkgs.is_empty()
|
||||||
|
&& added_direct_pkgs.is_empty()
|
||||||
|
&& removed_direct_pkgs.is_empty();
|
||||||
|
|
||||||
|
if no_changes {
|
||||||
|
println!("{prefix} already up to date");
|
||||||
|
} else {
|
||||||
|
let mut change_signs = [
|
||||||
|
(!added_pkgs.is_empty()).then(|| format!("+{}", added_pkgs.len()).green().to_string()),
|
||||||
|
(!removed_pkgs.is_empty())
|
||||||
|
.then(|| format!("-{}", removed_pkgs.len()).red().to_string()),
|
||||||
|
]
|
||||||
|
.into_iter()
|
||||||
|
.flatten()
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
.join(" ");
|
||||||
|
|
||||||
|
let changes_empty = change_signs.is_empty();
|
||||||
|
if changes_empty {
|
||||||
|
change_signs = "(no changes)".dimmed().to_string();
|
||||||
|
}
|
||||||
|
|
||||||
|
println!("{prefix} {change_signs}");
|
||||||
|
|
||||||
|
if !changes_empty {
|
||||||
|
println!(
|
||||||
|
"{}{}",
|
||||||
|
"+".repeat(added_pkgs.len()).green(),
|
||||||
|
"-".repeat(removed_pkgs.len()).red()
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
let dependency_groups = added_direct_pkgs
|
||||||
|
.iter()
|
||||||
|
.map(|(key, node)| (true, key, node))
|
||||||
|
.chain(
|
||||||
|
removed_direct_pkgs
|
||||||
|
.iter()
|
||||||
|
.map(|(key, node)| (false, key, node)),
|
||||||
|
)
|
||||||
|
.filter_map(|(added, key, node)| {
|
||||||
|
node.direct.as_ref().map(|(_, _, ty)| (added, key, ty))
|
||||||
|
})
|
||||||
|
.fold(
|
||||||
|
BTreeMap::<DependencyType, BTreeSet<_>>::new(),
|
||||||
|
|mut map, (added, key, &ty)| {
|
||||||
|
map.entry(ty).or_default().insert((key, added));
|
||||||
|
map
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
for (ty, set) in dependency_groups {
|
||||||
|
println!();
|
||||||
|
|
||||||
|
let ty_name = match ty {
|
||||||
|
DependencyType::Standard => "dependencies",
|
||||||
|
DependencyType::Peer => "peer_dependencies",
|
||||||
|
DependencyType::Dev => "dev_dependencies",
|
||||||
|
};
|
||||||
|
println!("{}", format!("{ty_name}:").yellow().bold());
|
||||||
|
|
||||||
|
for ((name, version), added) in set {
|
||||||
|
println!(
|
||||||
|
"{} {} {}",
|
||||||
|
if added { "+".green() } else { "-".red() },
|
||||||
|
name,
|
||||||
|
version.to_string().dimmed()
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
println!();
|
||||||
|
}
|
||||||
|
}
|
|
@ -15,7 +15,6 @@ use std::{
|
||||||
future::Future,
|
future::Future,
|
||||||
path::PathBuf,
|
path::PathBuf,
|
||||||
str::FromStr,
|
str::FromStr,
|
||||||
time::Duration,
|
|
||||||
};
|
};
|
||||||
use tokio::pin;
|
use tokio::pin;
|
||||||
use tracing::instrument;
|
use tracing::instrument;
|
||||||
|
@ -24,6 +23,8 @@ pub mod auth;
|
||||||
pub mod commands;
|
pub mod commands;
|
||||||
pub mod config;
|
pub mod config;
|
||||||
pub mod files;
|
pub mod files;
|
||||||
|
pub mod install;
|
||||||
|
pub mod reporters;
|
||||||
#[cfg(feature = "version-management")]
|
#[cfg(feature = "version-management")]
|
||||||
pub mod version;
|
pub mod version;
|
||||||
|
|
||||||
|
@ -193,39 +194,6 @@ pub fn parse_gix_url(s: &str) -> Result<gix::Url, gix::url::parse::Error> {
|
||||||
s.try_into()
|
s.try_into()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn progress_bar<E: std::error::Error + Into<anyhow::Error>>(
|
|
||||||
len: u64,
|
|
||||||
mut rx: tokio::sync::mpsc::Receiver<Result<String, E>>,
|
|
||||||
prefix: String,
|
|
||||||
progress_msg: String,
|
|
||||||
finish_msg: String,
|
|
||||||
) -> anyhow::Result<()> {
|
|
||||||
let bar = indicatif::ProgressBar::new(len)
|
|
||||||
.with_style(
|
|
||||||
indicatif::ProgressStyle::default_bar()
|
|
||||||
.template("{prefix}[{elapsed_precise}] {bar:40.208/166} {pos}/{len} {msg}")?
|
|
||||||
.progress_chars("█▓▒░ "),
|
|
||||||
)
|
|
||||||
.with_prefix(prefix)
|
|
||||||
.with_message(progress_msg);
|
|
||||||
bar.enable_steady_tick(Duration::from_millis(100));
|
|
||||||
|
|
||||||
while let Some(result) = rx.recv().await {
|
|
||||||
bar.inc(1);
|
|
||||||
|
|
||||||
match result {
|
|
||||||
Ok(text) => {
|
|
||||||
bar.set_message(text);
|
|
||||||
}
|
|
||||||
Err(e) => return Err(e.into()),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
bar.finish_with_message(finish_msg);
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn shift_project_dir(project: &Project, pkg_dir: PathBuf) -> Project {
|
pub fn shift_project_dir(project: &Project, pkg_dir: PathBuf) -> Project {
|
||||||
Project::new(
|
Project::new(
|
||||||
pkg_dir,
|
pkg_dir,
|
||||||
|
|
213
src/cli/reporters.rs
Normal file
213
src/cli/reporters.rs
Normal file
|
@ -0,0 +1,213 @@
|
||||||
|
//! Progress reporters for the CLI
|
||||||
|
|
||||||
|
use std::{
|
||||||
|
future::Future,
|
||||||
|
io::{Stdout, Write},
|
||||||
|
sync::{Arc, Mutex, Once, OnceLock},
|
||||||
|
time::Duration,
|
||||||
|
};
|
||||||
|
|
||||||
|
use colored::Colorize;
|
||||||
|
use indicatif::{MultiProgress, ProgressBar, ProgressStyle};
|
||||||
|
use pesde::reporters::{
|
||||||
|
DownloadProgressReporter, DownloadsReporter, PatchProgressReporter, PatchesReporter,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const TICK_CHARS: &str = "⣷⣯⣟⡿⢿⣻⣽⣾";
|
||||||
|
|
||||||
|
pub fn root_progress_style() -> ProgressStyle {
|
||||||
|
ProgressStyle::with_template("{prefix:.dim}{msg:>8.214/yellow} {spinner} [{elapsed_precise}]")
|
||||||
|
.unwrap()
|
||||||
|
.tick_chars(TICK_CHARS)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn root_progress_style_with_progress() -> ProgressStyle {
|
||||||
|
ProgressStyle::with_template(
|
||||||
|
"{prefix:.dim}{msg:>8.214/yellow} {spinner} [{elapsed_precise}] {bar:20} {pos}/{len}",
|
||||||
|
)
|
||||||
|
.unwrap()
|
||||||
|
.tick_chars(TICK_CHARS)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_with_reporter_and_writer<W, F, R, Fut>(writer: W, f: F) -> R
|
||||||
|
where
|
||||||
|
W: Write + Send + Sync + 'static,
|
||||||
|
F: FnOnce(MultiProgress, ProgressBar, Arc<CliReporter<W>>) -> Fut,
|
||||||
|
Fut: Future<Output = R>,
|
||||||
|
{
|
||||||
|
let multi_progress = MultiProgress::new();
|
||||||
|
crate::PROGRESS_BARS
|
||||||
|
.lock()
|
||||||
|
.unwrap()
|
||||||
|
.replace(multi_progress.clone());
|
||||||
|
|
||||||
|
let root_progress = multi_progress.add(ProgressBar::new(0));
|
||||||
|
root_progress.set_style(root_progress_style());
|
||||||
|
root_progress.enable_steady_tick(Duration::from_millis(100));
|
||||||
|
|
||||||
|
let reporter = Arc::new(CliReporter::with_writer(
|
||||||
|
writer,
|
||||||
|
multi_progress.clone(),
|
||||||
|
root_progress.clone(),
|
||||||
|
));
|
||||||
|
let result = f(multi_progress.clone(), root_progress.clone(), reporter).await;
|
||||||
|
|
||||||
|
root_progress.finish();
|
||||||
|
multi_progress.clear().unwrap();
|
||||||
|
crate::PROGRESS_BARS.lock().unwrap().take();
|
||||||
|
|
||||||
|
result
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_with_reporter<F, R, Fut>(f: F) -> R
|
||||||
|
where
|
||||||
|
F: FnOnce(MultiProgress, ProgressBar, Arc<CliReporter<Stdout>>) -> Fut,
|
||||||
|
Fut: Future<Output = R>,
|
||||||
|
{
|
||||||
|
run_with_reporter_and_writer(std::io::stdout(), f).await
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct CliReporter<W = Stdout> {
|
||||||
|
writer: Mutex<W>,
|
||||||
|
child_style: ProgressStyle,
|
||||||
|
child_style_with_bytes: ProgressStyle,
|
||||||
|
child_style_with_bytes_without_total: ProgressStyle,
|
||||||
|
multi_progress: MultiProgress,
|
||||||
|
root_progress: ProgressBar,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<W> CliReporter<W> {
|
||||||
|
pub fn with_writer(
|
||||||
|
writer: W,
|
||||||
|
multi_progress: MultiProgress,
|
||||||
|
root_progress: ProgressBar,
|
||||||
|
) -> Self {
|
||||||
|
Self {
|
||||||
|
writer: Mutex::new(writer),
|
||||||
|
child_style: ProgressStyle::with_template(&"{msg}".dimmed().to_string()).unwrap(),
|
||||||
|
child_style_with_bytes: ProgressStyle::with_template(
|
||||||
|
&"{msg} {bytes}/{total_bytes}".dimmed().to_string(),
|
||||||
|
)
|
||||||
|
.unwrap(),
|
||||||
|
child_style_with_bytes_without_total: ProgressStyle::with_template(
|
||||||
|
&"{msg} {bytes}".dimmed().to_string(),
|
||||||
|
)
|
||||||
|
.unwrap(),
|
||||||
|
multi_progress,
|
||||||
|
root_progress,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct CliDownloadProgressReporter<'a, W> {
|
||||||
|
root_reporter: &'a CliReporter<W>,
|
||||||
|
name: String,
|
||||||
|
progress: OnceLock<ProgressBar>,
|
||||||
|
set_progress: Once,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a, W: Write + Send + Sync + 'static> DownloadsReporter<'a> for CliReporter<W> {
|
||||||
|
type DownloadProgressReporter = CliDownloadProgressReporter<'a, W>;
|
||||||
|
|
||||||
|
fn report_download<'b>(&'a self, name: &'b str) -> Self::DownloadProgressReporter {
|
||||||
|
self.root_progress.inc_length(1);
|
||||||
|
|
||||||
|
CliDownloadProgressReporter {
|
||||||
|
root_reporter: self,
|
||||||
|
name: name.to_string(),
|
||||||
|
progress: OnceLock::new(),
|
||||||
|
set_progress: Once::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<W: Write + Send + Sync + 'static> DownloadProgressReporter
|
||||||
|
for CliDownloadProgressReporter<'_, W>
|
||||||
|
{
|
||||||
|
fn report_start(&self) {
|
||||||
|
let progress = self.root_reporter.multi_progress.add(ProgressBar::new(0));
|
||||||
|
progress.set_style(self.root_reporter.child_style.clone());
|
||||||
|
progress.set_message(format!("- {}", self.name));
|
||||||
|
|
||||||
|
self.progress
|
||||||
|
.set(progress)
|
||||||
|
.expect("report_start called more than once");
|
||||||
|
}
|
||||||
|
|
||||||
|
fn report_progress(&self, total: u64, len: u64) {
|
||||||
|
if let Some(progress) = self.progress.get() {
|
||||||
|
progress.set_length(total);
|
||||||
|
progress.set_position(len);
|
||||||
|
|
||||||
|
self.set_progress.call_once(|| {
|
||||||
|
if total > 0 {
|
||||||
|
progress.set_style(self.root_reporter.child_style_with_bytes.clone());
|
||||||
|
} else {
|
||||||
|
progress.set_style(
|
||||||
|
self.root_reporter
|
||||||
|
.child_style_with_bytes_without_total
|
||||||
|
.clone(),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn report_done(&self) {
|
||||||
|
if let Some(progress) = self.progress.get() {
|
||||||
|
if progress.is_hidden() {
|
||||||
|
writeln!(
|
||||||
|
self.root_reporter.writer.lock().unwrap(),
|
||||||
|
"downloaded {}",
|
||||||
|
self.name
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
|
|
||||||
|
progress.finish();
|
||||||
|
self.root_reporter.multi_progress.remove(progress);
|
||||||
|
self.root_reporter.root_progress.inc(1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct CliPatchProgressReporter<'a, W> {
|
||||||
|
root_reporter: &'a CliReporter<W>,
|
||||||
|
name: String,
|
||||||
|
progress: ProgressBar,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a, W: Write + Send + Sync + 'static> PatchesReporter<'a> for CliReporter<W> {
|
||||||
|
type PatchProgressReporter = CliPatchProgressReporter<'a, W>;
|
||||||
|
|
||||||
|
fn report_patch<'b>(&'a self, name: &'b str) -> Self::PatchProgressReporter {
|
||||||
|
let progress = self.multi_progress.add(ProgressBar::new(0));
|
||||||
|
progress.set_style(self.child_style.clone());
|
||||||
|
progress.set_message(format!("- {name}"));
|
||||||
|
|
||||||
|
self.root_progress.inc_length(1);
|
||||||
|
|
||||||
|
CliPatchProgressReporter {
|
||||||
|
root_reporter: self,
|
||||||
|
name: name.to_string(),
|
||||||
|
progress,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<W: Write + Send + Sync + 'static> PatchProgressReporter for CliPatchProgressReporter<'_, W> {
|
||||||
|
fn report_done(&self) {
|
||||||
|
if self.progress.is_hidden() {
|
||||||
|
writeln!(
|
||||||
|
self.root_reporter.writer.lock().unwrap(),
|
||||||
|
"patched {}",
|
||||||
|
self.name
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
|
|
||||||
|
self.progress.finish();
|
||||||
|
self.root_reporter.multi_progress.remove(&self.progress);
|
||||||
|
self.root_reporter.root_progress.inc(1);
|
||||||
|
}
|
||||||
|
}
|
229
src/download.rs
229
src/download.rs
|
@ -1,50 +1,132 @@
|
||||||
use crate::{
|
use crate::{
|
||||||
lockfile::{DependencyGraph, DownloadedDependencyGraphNode, DownloadedGraph},
|
lockfile::{DependencyGraph, DownloadedDependencyGraphNode},
|
||||||
manifest::DependencyType,
|
manifest::DependencyType,
|
||||||
|
names::PackageNames,
|
||||||
refresh_sources,
|
refresh_sources,
|
||||||
|
reporters::{DownloadProgressReporter, DownloadsReporter},
|
||||||
source::{
|
source::{
|
||||||
traits::{PackageRef, PackageSource},
|
traits::{PackageRef, PackageSource},
|
||||||
|
version_id::VersionId,
|
||||||
PackageSources,
|
PackageSources,
|
||||||
},
|
},
|
||||||
Project, PACKAGES_CONTAINER_NAME,
|
Project, PACKAGES_CONTAINER_NAME,
|
||||||
};
|
};
|
||||||
|
use async_stream::try_stream;
|
||||||
use fs_err::tokio as fs;
|
use fs_err::tokio as fs;
|
||||||
use std::{
|
use futures::Stream;
|
||||||
collections::HashSet,
|
use std::{collections::HashSet, num::NonZeroUsize, sync::Arc};
|
||||||
sync::{Arc, Mutex},
|
use tokio::{sync::Semaphore, task::JoinSet};
|
||||||
};
|
|
||||||
use tracing::{instrument, Instrument};
|
use tracing::{instrument, Instrument};
|
||||||
|
|
||||||
type MultithreadedGraph = Arc<Mutex<DownloadedGraph>>;
|
/// Options for downloading.
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct DownloadGraphOptions<Reporter> {
|
||||||
|
/// The reqwest client.
|
||||||
|
pub reqwest: reqwest::Client,
|
||||||
|
/// The downloads reporter.
|
||||||
|
pub reporter: Option<Arc<Reporter>>,
|
||||||
|
/// Whether to skip dev dependencies.
|
||||||
|
pub prod: bool,
|
||||||
|
/// Whether to write the downloaded packages to disk.
|
||||||
|
pub write: bool,
|
||||||
|
/// Whether to download Wally packages.
|
||||||
|
pub wally: bool,
|
||||||
|
/// The max number of concurrent network requests.
|
||||||
|
pub network_concurrency: NonZeroUsize,
|
||||||
|
}
|
||||||
|
|
||||||
pub(crate) type MultithreadDownloadJob = (
|
impl<Reporter> DownloadGraphOptions<Reporter>
|
||||||
tokio::sync::mpsc::Receiver<Result<String, errors::DownloadGraphError>>,
|
where
|
||||||
MultithreadedGraph,
|
Reporter: for<'a> DownloadsReporter<'a> + Send + Sync + 'static,
|
||||||
);
|
{
|
||||||
|
/// Creates a new download options with the given reqwest client and reporter.
|
||||||
|
pub fn new(reqwest: reqwest::Client) -> Self {
|
||||||
|
Self {
|
||||||
|
reqwest,
|
||||||
|
reporter: None,
|
||||||
|
prod: false,
|
||||||
|
write: false,
|
||||||
|
wally: false,
|
||||||
|
network_concurrency: NonZeroUsize::new(16).unwrap(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the downloads reporter.
|
||||||
|
pub fn reporter(mut self, reporter: impl Into<Arc<Reporter>>) -> Self {
|
||||||
|
self.reporter.replace(reporter.into());
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets whether to skip dev dependencies.
|
||||||
|
pub fn prod(mut self, prod: bool) -> Self {
|
||||||
|
self.prod = prod;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets whether to write the downloaded packages to disk.
|
||||||
|
pub fn write(mut self, write: bool) -> Self {
|
||||||
|
self.write = write;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets whether to download Wally packages.
|
||||||
|
pub fn wally(mut self, wally: bool) -> Self {
|
||||||
|
self.wally = wally;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the max number of concurrent network requests.
|
||||||
|
pub fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self {
|
||||||
|
self.network_concurrency = network_concurrency;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<Reporter> Clone for DownloadGraphOptions<Reporter> {
|
||||||
|
fn clone(&self) -> Self {
|
||||||
|
Self {
|
||||||
|
reqwest: self.reqwest.clone(),
|
||||||
|
reporter: self.reporter.clone(),
|
||||||
|
prod: self.prod,
|
||||||
|
write: self.write,
|
||||||
|
wally: self.wally,
|
||||||
|
network_concurrency: self.network_concurrency,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl Project {
|
impl Project {
|
||||||
/// Downloads a graph of dependencies
|
/// Downloads a graph of dependencies.
|
||||||
#[instrument(skip(self, graph, refreshed_sources, reqwest), level = "debug")]
|
#[instrument(skip_all, fields(prod = options.prod, wally = options.wally, write = options.write), level = "debug")]
|
||||||
pub async fn download_graph(
|
pub async fn download_graph<Reporter>(
|
||||||
&self,
|
&self,
|
||||||
graph: &DependencyGraph,
|
graph: &DependencyGraph,
|
||||||
refreshed_sources: &mut HashSet<PackageSources>,
|
refreshed_sources: &mut HashSet<PackageSources>,
|
||||||
reqwest: &reqwest::Client,
|
options: DownloadGraphOptions<Reporter>,
|
||||||
prod: bool,
|
) -> Result<
|
||||||
write: bool,
|
impl Stream<
|
||||||
wally: bool,
|
Item = Result<
|
||||||
) -> Result<MultithreadDownloadJob, errors::DownloadGraphError> {
|
(DownloadedDependencyGraphNode, PackageNames, VersionId),
|
||||||
|
errors::DownloadGraphError,
|
||||||
|
>,
|
||||||
|
>,
|
||||||
|
errors::DownloadGraphError,
|
||||||
|
>
|
||||||
|
where
|
||||||
|
Reporter: for<'a> DownloadsReporter<'a> + Send + Sync + 'static,
|
||||||
|
{
|
||||||
|
let DownloadGraphOptions {
|
||||||
|
reqwest,
|
||||||
|
reporter,
|
||||||
|
prod,
|
||||||
|
write,
|
||||||
|
wally,
|
||||||
|
network_concurrency,
|
||||||
|
} = options;
|
||||||
|
|
||||||
let manifest = self.deser_manifest().await?;
|
let manifest = self.deser_manifest().await?;
|
||||||
let manifest_target_kind = manifest.target.kind();
|
let manifest_target_kind = manifest.target.kind();
|
||||||
let downloaded_graph: MultithreadedGraph = Arc::new(Mutex::new(Default::default()));
|
let project = Arc::new(self.clone());
|
||||||
|
|
||||||
let (tx, rx) = tokio::sync::mpsc::channel(
|
|
||||||
graph
|
|
||||||
.iter()
|
|
||||||
.map(|(_, versions)| versions.len())
|
|
||||||
.sum::<usize>()
|
|
||||||
.max(1),
|
|
||||||
);
|
|
||||||
|
|
||||||
refresh_sources(
|
refresh_sources(
|
||||||
self,
|
self,
|
||||||
|
@ -56,7 +138,8 @@ impl Project {
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
|
|
||||||
let project = Arc::new(self.clone());
|
let mut tasks = JoinSet::<Result<_, errors::DownloadGraphError>>::new();
|
||||||
|
let semaphore = Arc::new(Semaphore::new(network_concurrency.get()));
|
||||||
|
|
||||||
for (name, versions) in graph {
|
for (name, versions) in graph {
|
||||||
for (version_id, node) in versions {
|
for (version_id, node) in versions {
|
||||||
|
@ -65,8 +148,6 @@ impl Project {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
let tx = tx.clone();
|
|
||||||
|
|
||||||
let name = name.clone();
|
let name = name.clone();
|
||||||
let version_id = version_id.clone();
|
let version_id = version_id.clone();
|
||||||
let node = node.clone();
|
let node = node.clone();
|
||||||
|
@ -79,14 +160,24 @@ impl Project {
|
||||||
|
|
||||||
let project = project.clone();
|
let project = project.clone();
|
||||||
let reqwest = reqwest.clone();
|
let reqwest = reqwest.clone();
|
||||||
let downloaded_graph = downloaded_graph.clone();
|
let reporter = reporter.clone();
|
||||||
|
let package_dir = project.package_dir().to_path_buf();
|
||||||
|
let semaphore = semaphore.clone();
|
||||||
|
|
||||||
let package_dir = self.package_dir().to_path_buf();
|
tasks.spawn(
|
||||||
|
|
||||||
tokio::spawn(
|
|
||||||
async move {
|
async move {
|
||||||
let source = node.pkg_ref.source();
|
let display_name = format!("{name}@{version_id}");
|
||||||
|
let progress_reporter = reporter
|
||||||
|
.as_deref()
|
||||||
|
.map(|reporter| reporter.report_download(&display_name));
|
||||||
|
|
||||||
|
let _permit = semaphore.acquire().await;
|
||||||
|
|
||||||
|
if let Some(ref progress_reporter) = progress_reporter {
|
||||||
|
progress_reporter.report_start();
|
||||||
|
}
|
||||||
|
|
||||||
|
let source = node.pkg_ref.source();
|
||||||
let container_folder = node.container_folder(
|
let container_folder = node.container_folder(
|
||||||
&package_dir
|
&package_dir
|
||||||
.join(manifest_target_kind.packages_folder(version_id.target()))
|
.join(manifest_target_kind.packages_folder(version_id.target()))
|
||||||
|
@ -95,42 +186,37 @@ impl Project {
|
||||||
version_id.version(),
|
version_id.version(),
|
||||||
);
|
);
|
||||||
|
|
||||||
match fs::create_dir_all(&container_folder).await {
|
fs::create_dir_all(&container_folder).await?;
|
||||||
Ok(_) => {}
|
|
||||||
Err(e) => {
|
|
||||||
tx.send(Err(errors::DownloadGraphError::Io(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let project = project.clone();
|
let project = project.clone();
|
||||||
|
|
||||||
tracing::debug!("downloading");
|
tracing::debug!("downloading");
|
||||||
|
|
||||||
let (fs, target) =
|
let (fs, target) = match progress_reporter {
|
||||||
match source.download(&node.pkg_ref, &project, &reqwest).await {
|
Some(progress_reporter) => {
|
||||||
Ok(target) => target,
|
source
|
||||||
Err(e) => {
|
.download(
|
||||||
tx.send(Err(Box::new(e).into())).await.unwrap();
|
&node.pkg_ref,
|
||||||
return;
|
&project,
|
||||||
}
|
&reqwest,
|
||||||
};
|
Arc::new(progress_reporter),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
source
|
||||||
|
.download(&node.pkg_ref, &project, &reqwest, Arc::new(()))
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
}
|
||||||
|
.map_err(Box::new)?;
|
||||||
|
|
||||||
tracing::debug!("downloaded");
|
tracing::debug!("downloaded");
|
||||||
|
|
||||||
if write {
|
if write {
|
||||||
if !prod || node.resolved_ty != DependencyType::Dev {
|
if !prod || node.resolved_ty != DependencyType::Dev {
|
||||||
match fs.write_to(container_folder, project.cas_dir(), true).await {
|
fs.write_to(container_folder, project.cas_dir(), true)
|
||||||
Ok(_) => {}
|
.await?;
|
||||||
Err(e) => {
|
|
||||||
tx.send(Err(errors::DownloadGraphError::WriteFailed(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
} else {
|
} else {
|
||||||
tracing::debug!(
|
tracing::debug!(
|
||||||
"skipping write to disk, dev dependency in prod mode"
|
"skipping write to disk, dev dependency in prod mode"
|
||||||
|
@ -138,24 +224,21 @@ impl Project {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let display_name = format!("{name}@{version_id}");
|
let downloaded_node = DownloadedDependencyGraphNode { node, target };
|
||||||
|
Ok((downloaded_node, name, version_id))
|
||||||
{
|
|
||||||
let mut downloaded_graph = downloaded_graph.lock().unwrap();
|
|
||||||
downloaded_graph
|
|
||||||
.entry(name)
|
|
||||||
.or_default()
|
|
||||||
.insert(version_id, DownloadedDependencyGraphNode { node, target });
|
|
||||||
}
|
|
||||||
|
|
||||||
tx.send(Ok(display_name)).await.unwrap();
|
|
||||||
}
|
}
|
||||||
.instrument(span),
|
.instrument(span),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok((rx, downloaded_graph))
|
let stream = try_stream! {
|
||||||
|
while let Some(res) = tasks.join_next().await {
|
||||||
|
yield res.unwrap()?;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
Ok(stream)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,14 +1,18 @@
|
||||||
use crate::{
|
use crate::{
|
||||||
|
download::DownloadGraphOptions,
|
||||||
lockfile::{DependencyGraph, DownloadedGraph},
|
lockfile::{DependencyGraph, DownloadedGraph},
|
||||||
manifest::DependencyType,
|
manifest::DependencyType,
|
||||||
|
reporters::DownloadsReporter,
|
||||||
source::PackageSources,
|
source::PackageSources,
|
||||||
Project,
|
Project,
|
||||||
};
|
};
|
||||||
use futures::FutureExt;
|
use futures::TryStreamExt;
|
||||||
use std::{
|
use std::{
|
||||||
collections::HashSet,
|
collections::HashSet,
|
||||||
future::Future,
|
convert::Infallible,
|
||||||
sync::{Arc, Mutex as StdMutex},
|
future::{self, Future},
|
||||||
|
num::NonZeroUsize,
|
||||||
|
sync::Arc,
|
||||||
};
|
};
|
||||||
use tokio::sync::Mutex;
|
use tokio::sync::Mutex;
|
||||||
use tracing::{instrument, Instrument};
|
use tracing::{instrument, Instrument};
|
||||||
|
@ -38,118 +42,242 @@ pub fn filter_graph(graph: &DownloadedGraph, prod: bool) -> DownloadedGraph {
|
||||||
pub type DownloadAndLinkReceiver =
|
pub type DownloadAndLinkReceiver =
|
||||||
tokio::sync::mpsc::Receiver<Result<String, crate::download::errors::DownloadGraphError>>;
|
tokio::sync::mpsc::Receiver<Result<String, crate::download::errors::DownloadGraphError>>;
|
||||||
|
|
||||||
|
/// Hooks to perform actions after certain events during download and linking.
|
||||||
|
#[allow(unused_variables)]
|
||||||
|
pub trait DownloadAndLinkHooks {
|
||||||
|
/// The error type for the hooks.
|
||||||
|
type Error: std::error::Error + Send + Sync + 'static;
|
||||||
|
|
||||||
|
/// Called after scripts have been downloaded. The `downloaded_graph`
|
||||||
|
/// contains all downloaded packages.
|
||||||
|
fn on_scripts_downloaded(
|
||||||
|
&self,
|
||||||
|
downloaded_graph: &DownloadedGraph,
|
||||||
|
) -> impl Future<Output = Result<(), Self::Error>> + Send {
|
||||||
|
future::ready(Ok(()))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Called after binary dependencies have been downloaded. The
|
||||||
|
/// `downloaded_graph` contains all downloaded packages.
|
||||||
|
fn on_bins_downloaded(
|
||||||
|
&self,
|
||||||
|
downloaded_graph: &DownloadedGraph,
|
||||||
|
) -> impl Future<Output = Result<(), Self::Error>> + Send {
|
||||||
|
future::ready(Ok(()))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Called after all dependencies have been downloaded. The
|
||||||
|
/// `downloaded_graph` contains all downloaded packages.
|
||||||
|
fn on_all_downloaded(
|
||||||
|
&self,
|
||||||
|
downloaded_graph: &DownloadedGraph,
|
||||||
|
) -> impl Future<Output = Result<(), Self::Error>> + Send {
|
||||||
|
future::ready(Ok(()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DownloadAndLinkHooks for () {
|
||||||
|
type Error = Infallible;
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Options for downloading and linking.
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct DownloadAndLinkOptions<Reporter = (), Hooks = ()> {
|
||||||
|
/// The reqwest client.
|
||||||
|
pub reqwest: reqwest::Client,
|
||||||
|
/// The downloads reporter.
|
||||||
|
pub reporter: Option<Arc<Reporter>>,
|
||||||
|
/// The download and link hooks.
|
||||||
|
pub hooks: Option<Arc<Hooks>>,
|
||||||
|
/// The refreshed sources.
|
||||||
|
pub refreshed_sources: Arc<Mutex<HashSet<PackageSources>>>,
|
||||||
|
/// Whether to skip dev dependencies.
|
||||||
|
pub prod: bool,
|
||||||
|
/// Whether to write the downloaded packages to disk.
|
||||||
|
pub write: bool,
|
||||||
|
/// The max number of concurrent network requests.
|
||||||
|
pub network_concurrency: NonZeroUsize,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<Reporter, Hooks> DownloadAndLinkOptions<Reporter, Hooks>
|
||||||
|
where
|
||||||
|
Reporter: for<'a> DownloadsReporter<'a> + Send + Sync + 'static,
|
||||||
|
Hooks: DownloadAndLinkHooks + Send + Sync + 'static,
|
||||||
|
{
|
||||||
|
/// Creates a new download options with the given reqwest client and reporter.
|
||||||
|
pub fn new(reqwest: reqwest::Client) -> Self {
|
||||||
|
Self {
|
||||||
|
reqwest,
|
||||||
|
reporter: None,
|
||||||
|
hooks: None,
|
||||||
|
refreshed_sources: Default::default(),
|
||||||
|
prod: false,
|
||||||
|
write: true,
|
||||||
|
network_concurrency: NonZeroUsize::new(16).unwrap(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the downloads reporter.
|
||||||
|
pub fn reporter(mut self, reporter: impl Into<Arc<Reporter>>) -> Self {
|
||||||
|
self.reporter.replace(reporter.into());
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the download and link hooks.
|
||||||
|
pub fn hooks(mut self, hooks: impl Into<Arc<Hooks>>) -> Self {
|
||||||
|
self.hooks.replace(hooks.into());
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the refreshed sources.
|
||||||
|
pub fn refreshed_sources(
|
||||||
|
mut self,
|
||||||
|
refreshed_sources: impl Into<Arc<Mutex<HashSet<PackageSources>>>>,
|
||||||
|
) -> Self {
|
||||||
|
self.refreshed_sources = refreshed_sources.into();
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets whether to skip dev dependencies.
|
||||||
|
pub fn prod(mut self, prod: bool) -> Self {
|
||||||
|
self.prod = prod;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets whether to write the downloaded packages to disk.
|
||||||
|
pub fn write(mut self, write: bool) -> Self {
|
||||||
|
self.write = write;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sets the max number of concurrent network requests.
|
||||||
|
pub fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self {
|
||||||
|
self.network_concurrency = network_concurrency;
|
||||||
|
self
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Clone for DownloadAndLinkOptions {
|
||||||
|
fn clone(&self) -> Self {
|
||||||
|
Self {
|
||||||
|
reqwest: self.reqwest.clone(),
|
||||||
|
reporter: self.reporter.clone(),
|
||||||
|
hooks: self.hooks.clone(),
|
||||||
|
refreshed_sources: self.refreshed_sources.clone(),
|
||||||
|
prod: self.prod,
|
||||||
|
write: self.write,
|
||||||
|
network_concurrency: self.network_concurrency,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl Project {
|
impl Project {
|
||||||
/// Downloads a graph of dependencies and links them in the correct order
|
/// Downloads a graph of dependencies and links them in the correct order
|
||||||
#[instrument(
|
#[instrument(skip_all, fields(prod = options.prod, write = options.write), level = "debug")]
|
||||||
skip(self, graph, refreshed_sources, reqwest, pesde_cb),
|
pub async fn download_and_link<Reporter, Hooks>(
|
||||||
level = "debug"
|
|
||||||
)]
|
|
||||||
pub async fn download_and_link<
|
|
||||||
F: FnOnce(&Arc<DownloadedGraph>) -> R + Send + 'static,
|
|
||||||
R: Future<Output = Result<(), E>> + Send,
|
|
||||||
E: Send + Sync + 'static,
|
|
||||||
>(
|
|
||||||
&self,
|
&self,
|
||||||
graph: &Arc<DependencyGraph>,
|
graph: &Arc<DependencyGraph>,
|
||||||
refreshed_sources: &Arc<Mutex<HashSet<PackageSources>>>,
|
options: DownloadAndLinkOptions<Reporter, Hooks>,
|
||||||
reqwest: &reqwest::Client,
|
) -> Result<DownloadedGraph, errors::DownloadAndLinkError<Hooks::Error>>
|
||||||
prod: bool,
|
where
|
||||||
write: bool,
|
Reporter: for<'a> DownloadsReporter<'a> + 'static,
|
||||||
pesde_cb: F,
|
Hooks: DownloadAndLinkHooks + 'static,
|
||||||
) -> Result<
|
{
|
||||||
(
|
let DownloadAndLinkOptions {
|
||||||
DownloadAndLinkReceiver,
|
reqwest,
|
||||||
impl Future<Output = Result<DownloadedGraph, errors::DownloadAndLinkError<E>>>,
|
reporter,
|
||||||
),
|
hooks,
|
||||||
errors::DownloadAndLinkError<E>,
|
refreshed_sources,
|
||||||
> {
|
prod,
|
||||||
let (tx, rx) = tokio::sync::mpsc::channel(
|
write,
|
||||||
graph
|
network_concurrency,
|
||||||
.iter()
|
} = options;
|
||||||
.map(|(_, versions)| versions.len())
|
|
||||||
.sum::<usize>()
|
|
||||||
.max(1),
|
|
||||||
);
|
|
||||||
let downloaded_graph = Arc::new(StdMutex::new(DownloadedGraph::default()));
|
|
||||||
|
|
||||||
let this = self.clone();
|
|
||||||
let graph = graph.clone();
|
let graph = graph.clone();
|
||||||
let reqwest = reqwest.clone();
|
let reqwest = reqwest.clone();
|
||||||
let refreshed_sources = refreshed_sources.clone();
|
|
||||||
|
|
||||||
Ok((
|
let mut refreshed_sources = refreshed_sources.lock().await;
|
||||||
rx,
|
let mut downloaded_graph = DownloadedGraph::new();
|
||||||
tokio::spawn(async move {
|
|
||||||
let mut refreshed_sources = refreshed_sources.lock().await;
|
|
||||||
|
|
||||||
// step 1. download pesde dependencies
|
let mut download_graph_options = DownloadGraphOptions::<Reporter>::new(reqwest.clone())
|
||||||
let (mut pesde_rx, pesde_graph) = this
|
.prod(prod)
|
||||||
.download_graph(&graph, &mut refreshed_sources, &reqwest, prod, write, false)
|
.write(write)
|
||||||
.instrument(tracing::debug_span!("download (pesde)"))
|
.network_concurrency(network_concurrency);
|
||||||
.await?;
|
|
||||||
|
|
||||||
while let Some(result) = pesde_rx.recv().await {
|
if let Some(reporter) = reporter {
|
||||||
tx.send(result).await.unwrap();
|
download_graph_options = download_graph_options.reporter(reporter.clone());
|
||||||
}
|
}
|
||||||
|
|
||||||
let pesde_graph = Arc::into_inner(pesde_graph).unwrap().into_inner().unwrap();
|
// step 1. download pesde dependencies
|
||||||
|
self.download_graph(
|
||||||
|
&graph,
|
||||||
|
&mut refreshed_sources,
|
||||||
|
download_graph_options.clone(),
|
||||||
|
)
|
||||||
|
.instrument(tracing::debug_span!("download (pesde)"))
|
||||||
|
.await?
|
||||||
|
.try_for_each(|(downloaded_node, name, version_id)| {
|
||||||
|
downloaded_graph
|
||||||
|
.entry(name)
|
||||||
|
.or_default()
|
||||||
|
.insert(version_id, downloaded_node);
|
||||||
|
|
||||||
// step 2. link pesde dependencies. do so without types
|
future::ready(Ok(()))
|
||||||
if write {
|
})
|
||||||
this.link_dependencies(&filter_graph(&pesde_graph, prod), false)
|
.await?;
|
||||||
.instrument(tracing::debug_span!("link (pesde)"))
|
|
||||||
.await?;
|
|
||||||
}
|
|
||||||
|
|
||||||
let pesde_graph = Arc::new(pesde_graph);
|
// step 2. link pesde dependencies. do so without types
|
||||||
|
if write {
|
||||||
|
self.link_dependencies(&filter_graph(&downloaded_graph, prod), false)
|
||||||
|
.instrument(tracing::debug_span!("link (pesde)"))
|
||||||
|
.await?;
|
||||||
|
}
|
||||||
|
|
||||||
pesde_cb(&pesde_graph)
|
if let Some(ref hooks) = hooks {
|
||||||
.await
|
hooks
|
||||||
.map_err(errors::DownloadAndLinkError::PesdeCallback)?;
|
.on_scripts_downloaded(&downloaded_graph)
|
||||||
|
.await
|
||||||
|
.map_err(errors::DownloadAndLinkError::Hook)?;
|
||||||
|
|
||||||
let pesde_graph = Arc::into_inner(pesde_graph).unwrap();
|
hooks
|
||||||
|
.on_bins_downloaded(&downloaded_graph)
|
||||||
|
.await
|
||||||
|
.map_err(errors::DownloadAndLinkError::Hook)?;
|
||||||
|
}
|
||||||
|
|
||||||
// step 3. download wally dependencies
|
// step 3. download wally dependencies
|
||||||
let (mut wally_rx, wally_graph) = this
|
self.download_graph(
|
||||||
.download_graph(&graph, &mut refreshed_sources, &reqwest, prod, write, true)
|
&graph,
|
||||||
.instrument(tracing::debug_span!("download (wally)"))
|
&mut refreshed_sources,
|
||||||
.await?;
|
download_graph_options.clone().wally(true),
|
||||||
|
)
|
||||||
|
.instrument(tracing::debug_span!("download (wally)"))
|
||||||
|
.await?
|
||||||
|
.try_for_each(|(downloaded_node, name, version_id)| {
|
||||||
|
downloaded_graph
|
||||||
|
.entry(name)
|
||||||
|
.or_default()
|
||||||
|
.insert(version_id, downloaded_node);
|
||||||
|
|
||||||
while let Some(result) = wally_rx.recv().await {
|
future::ready(Ok(()))
|
||||||
tx.send(result).await.unwrap();
|
})
|
||||||
}
|
.await?;
|
||||||
|
|
||||||
let wally_graph = Arc::into_inner(wally_graph).unwrap().into_inner().unwrap();
|
// step 4. link ALL dependencies. do so with types
|
||||||
|
if write {
|
||||||
|
self.link_dependencies(&filter_graph(&downloaded_graph, prod), true)
|
||||||
|
.instrument(tracing::debug_span!("link (all)"))
|
||||||
|
.await?;
|
||||||
|
}
|
||||||
|
|
||||||
{
|
if let Some(ref hooks) = hooks {
|
||||||
let mut downloaded_graph = downloaded_graph.lock().unwrap();
|
hooks
|
||||||
downloaded_graph.extend(pesde_graph);
|
.on_all_downloaded(&downloaded_graph)
|
||||||
for (name, versions) in wally_graph {
|
.await
|
||||||
for (version_id, node) in versions {
|
.map_err(errors::DownloadAndLinkError::Hook)?;
|
||||||
downloaded_graph
|
}
|
||||||
.entry(name.clone())
|
|
||||||
.or_default()
|
|
||||||
.insert(version_id, node);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let graph = Arc::into_inner(downloaded_graph)
|
Ok(downloaded_graph)
|
||||||
.unwrap()
|
|
||||||
.into_inner()
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
// step 4. link ALL dependencies. do so with types
|
|
||||||
if write {
|
|
||||||
this.link_dependencies(&filter_graph(&graph, prod), true)
|
|
||||||
.instrument(tracing::debug_span!("link (all)"))
|
|
||||||
.await?;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(graph)
|
|
||||||
})
|
|
||||||
.map(|r| r.unwrap()),
|
|
||||||
))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -170,7 +298,7 @@ pub mod errors {
|
||||||
Linking(#[from] crate::linking::errors::LinkingError),
|
Linking(#[from] crate::linking::errors::LinkingError),
|
||||||
|
|
||||||
/// An error occurred while executing the pesde callback
|
/// An error occurred while executing the pesde callback
|
||||||
#[error("error executing pesde callback")]
|
#[error("error executing hook")]
|
||||||
PesdeCallback(#[source] E),
|
Hook(#[source] E),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -35,6 +35,7 @@ pub mod names;
|
||||||
/// Patching packages
|
/// Patching packages
|
||||||
#[cfg(feature = "patches")]
|
#[cfg(feature = "patches")]
|
||||||
pub mod patches;
|
pub mod patches;
|
||||||
|
pub mod reporters;
|
||||||
/// Resolving packages
|
/// Resolving packages
|
||||||
pub mod resolver;
|
pub mod resolver;
|
||||||
/// Running scripts
|
/// Running scripts
|
||||||
|
@ -182,9 +183,9 @@ impl Project {
|
||||||
#[instrument(skip(self, lockfile), level = "debug")]
|
#[instrument(skip(self, lockfile), level = "debug")]
|
||||||
pub async fn write_lockfile(
|
pub async fn write_lockfile(
|
||||||
&self,
|
&self,
|
||||||
lockfile: Lockfile,
|
lockfile: &Lockfile,
|
||||||
) -> Result<(), errors::LockfileWriteError> {
|
) -> Result<(), errors::LockfileWriteError> {
|
||||||
let string = toml::to_string(&lockfile)?;
|
let string = toml::to_string(lockfile)?;
|
||||||
fs::write(self.package_dir.join(LOCKFILE_FILE_NAME), string).await?;
|
fs::write(self.package_dir.join(LOCKFILE_FILE_NAME), string).await?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
56
src/main.rs
56
src/main.rs
|
@ -4,16 +4,18 @@ use crate::cli::{auth::get_tokens, display_err, home_dir, HOME_DIR};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use clap::{builder::styling::AnsiColor, Parser};
|
use clap::{builder::styling::AnsiColor, Parser};
|
||||||
use fs_err::tokio as fs;
|
use fs_err::tokio as fs;
|
||||||
|
use indicatif::MultiProgress;
|
||||||
use pesde::{matching_globs, AuthConfig, Project, MANIFEST_FILE_NAME};
|
use pesde::{matching_globs, AuthConfig, Project, MANIFEST_FILE_NAME};
|
||||||
use std::{
|
use std::{
|
||||||
collections::HashSet,
|
collections::HashSet,
|
||||||
|
io,
|
||||||
path::{Path, PathBuf},
|
path::{Path, PathBuf},
|
||||||
|
sync::Mutex,
|
||||||
};
|
};
|
||||||
use tempfile::NamedTempFile;
|
use tempfile::NamedTempFile;
|
||||||
use tracing::instrument;
|
use tracing::instrument;
|
||||||
use tracing_indicatif::{filter::IndicatifFilter, IndicatifLayer};
|
|
||||||
use tracing_subscriber::{
|
use tracing_subscriber::{
|
||||||
filter::LevelFilter, layer::SubscriberExt, util::SubscriberInitExt, EnvFilter, Layer,
|
filter::LevelFilter, fmt::MakeWriter, layer::SubscriberExt, util::SubscriberInitExt, EnvFilter,
|
||||||
};
|
};
|
||||||
|
|
||||||
mod cli;
|
mod cli;
|
||||||
|
@ -88,6 +90,50 @@ async fn get_linkable_dir(path: &Path) -> PathBuf {
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub static PROGRESS_BARS: Mutex<Option<MultiProgress>> = Mutex::new(None);
|
||||||
|
|
||||||
|
#[derive(Clone, Copy)]
|
||||||
|
pub struct IndicatifWriter;
|
||||||
|
|
||||||
|
impl IndicatifWriter {
|
||||||
|
fn suspend<F: FnOnce() -> R, R>(f: F) -> R {
|
||||||
|
match *PROGRESS_BARS.lock().unwrap() {
|
||||||
|
Some(ref progress_bars) => progress_bars.suspend(f),
|
||||||
|
None => f(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl io::Write for IndicatifWriter {
|
||||||
|
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
|
||||||
|
Self::suspend(|| io::stderr().write(buf))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn flush(&mut self) -> io::Result<()> {
|
||||||
|
Self::suspend(|| io::stderr().flush())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn write_vectored(&mut self, bufs: &[io::IoSlice<'_>]) -> io::Result<usize> {
|
||||||
|
Self::suspend(|| io::stderr().write_vectored(bufs))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn write_all(&mut self, buf: &[u8]) -> io::Result<()> {
|
||||||
|
Self::suspend(|| io::stderr().write_all(buf))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn write_fmt(&mut self, fmt: std::fmt::Arguments<'_>) -> io::Result<()> {
|
||||||
|
Self::suspend(|| io::stderr().write_fmt(fmt))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> MakeWriter<'a> for IndicatifWriter {
|
||||||
|
type Writer = IndicatifWriter;
|
||||||
|
|
||||||
|
fn make_writer(&'a self) -> Self::Writer {
|
||||||
|
*self
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
async fn run() -> anyhow::Result<()> {
|
async fn run() -> anyhow::Result<()> {
|
||||||
let cwd = std::env::current_dir().expect("failed to get current working directory");
|
let cwd = std::env::current_dir().expect("failed to get current working directory");
|
||||||
|
|
||||||
|
@ -133,8 +179,6 @@ async fn run() -> anyhow::Result<()> {
|
||||||
std::process::exit(status.code().unwrap());
|
std::process::exit(status.code().unwrap());
|
||||||
}
|
}
|
||||||
|
|
||||||
let indicatif_layer = IndicatifLayer::new().with_filter(IndicatifFilter::new(false));
|
|
||||||
|
|
||||||
let tracing_env_filter = EnvFilter::builder()
|
let tracing_env_filter = EnvFilter::builder()
|
||||||
.with_default_directive(LevelFilter::INFO.into())
|
.with_default_directive(LevelFilter::INFO.into())
|
||||||
.from_env_lossy()
|
.from_env_lossy()
|
||||||
|
@ -146,8 +190,7 @@ async fn run() -> anyhow::Result<()> {
|
||||||
.add_directive("hyper=info".parse().unwrap())
|
.add_directive("hyper=info".parse().unwrap())
|
||||||
.add_directive("h2=info".parse().unwrap());
|
.add_directive("h2=info".parse().unwrap());
|
||||||
|
|
||||||
let fmt_layer =
|
let fmt_layer = tracing_subscriber::fmt::layer().with_writer(IndicatifWriter);
|
||||||
tracing_subscriber::fmt::layer().with_writer(indicatif_layer.inner().get_stderr_writer());
|
|
||||||
|
|
||||||
#[cfg(debug_assertions)]
|
#[cfg(debug_assertions)]
|
||||||
let fmt_layer = fmt_layer.with_timer(tracing_subscriber::fmt::time::uptime());
|
let fmt_layer = fmt_layer.with_timer(tracing_subscriber::fmt::time::uptime());
|
||||||
|
@ -163,7 +206,6 @@ async fn run() -> anyhow::Result<()> {
|
||||||
tracing_subscriber::registry()
|
tracing_subscriber::registry()
|
||||||
.with(tracing_env_filter)
|
.with(tracing_env_filter)
|
||||||
.with(fmt_layer)
|
.with(fmt_layer)
|
||||||
.with(indicatif_layer)
|
|
||||||
.init();
|
.init();
|
||||||
|
|
||||||
let (project_root_dir, project_workspace_dir) = 'finder: {
|
let (project_root_dir, project_workspace_dir) = 'finder: {
|
||||||
|
|
|
@ -94,7 +94,7 @@ pub struct Manifest {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// A dependency type
|
/// A dependency type
|
||||||
#[derive(Serialize, Deserialize, Debug, Clone, Copy, PartialEq, Eq, Hash)]
|
#[derive(Serialize, Deserialize, Debug, Clone, Copy, PartialEq, Eq, Hash, PartialOrd, Ord)]
|
||||||
#[serde(rename_all = "snake_case")]
|
#[serde(rename_all = "snake_case")]
|
||||||
pub enum DependencyType {
|
pub enum DependencyType {
|
||||||
/// A standard dependency
|
/// A standard dependency
|
||||||
|
|
140
src/patches.rs
140
src/patches.rs
|
@ -1,8 +1,14 @@
|
||||||
use crate::{lockfile::DownloadedGraph, Project, MANIFEST_FILE_NAME, PACKAGES_CONTAINER_NAME};
|
use crate::{
|
||||||
|
lockfile::DownloadedGraph,
|
||||||
|
reporters::{PatchProgressReporter, PatchesReporter},
|
||||||
|
Project, MANIFEST_FILE_NAME, PACKAGES_CONTAINER_NAME,
|
||||||
|
};
|
||||||
use fs_err::tokio as fs;
|
use fs_err::tokio as fs;
|
||||||
|
use futures::TryFutureExt;
|
||||||
use git2::{ApplyLocation, Diff, DiffFormat, DiffLineType, Repository, Signature};
|
use git2::{ApplyLocation, Diff, DiffFormat, DiffLineType, Repository, Signature};
|
||||||
use relative_path::RelativePathBuf;
|
use relative_path::RelativePathBuf;
|
||||||
use std::path::Path;
|
use std::{path::Path, sync::Arc};
|
||||||
|
use tokio::task::JoinSet;
|
||||||
use tracing::instrument;
|
use tracing::instrument;
|
||||||
|
|
||||||
/// Set up a git repository for patches
|
/// Set up a git repository for patches
|
||||||
|
@ -70,28 +76,21 @@ pub fn create_patch<P: AsRef<Path>>(dir: P) -> Result<Vec<u8>, git2::Error> {
|
||||||
|
|
||||||
impl Project {
|
impl Project {
|
||||||
/// Apply patches to the project's dependencies
|
/// Apply patches to the project's dependencies
|
||||||
#[instrument(skip(self, graph), level = "debug")]
|
#[instrument(skip(self, graph, reporter), level = "debug")]
|
||||||
pub async fn apply_patches(
|
pub async fn apply_patches<Reporter>(
|
||||||
&self,
|
&self,
|
||||||
graph: &DownloadedGraph,
|
graph: &DownloadedGraph,
|
||||||
) -> Result<
|
reporter: Arc<Reporter>,
|
||||||
tokio::sync::mpsc::Receiver<Result<String, errors::ApplyPatchesError>>,
|
) -> Result<(), errors::ApplyPatchesError>
|
||||||
errors::ApplyPatchesError,
|
where
|
||||||
> {
|
Reporter: for<'a> PatchesReporter<'a> + Send + Sync + 'static,
|
||||||
|
{
|
||||||
let manifest = self.deser_manifest().await?;
|
let manifest = self.deser_manifest().await?;
|
||||||
let (tx, rx) = tokio::sync::mpsc::channel(
|
|
||||||
manifest
|
let mut tasks = JoinSet::<Result<_, errors::ApplyPatchesError>>::new();
|
||||||
.patches
|
|
||||||
.values()
|
|
||||||
.map(|v| v.len())
|
|
||||||
.sum::<usize>()
|
|
||||||
.max(1),
|
|
||||||
);
|
|
||||||
|
|
||||||
for (name, versions) in manifest.patches {
|
for (name, versions) in manifest.patches {
|
||||||
for (version_id, patch_path) in versions {
|
for (version_id, patch_path) in versions {
|
||||||
let tx = tx.clone();
|
|
||||||
|
|
||||||
let name = name.clone();
|
let name = name.clone();
|
||||||
let patch_path = patch_path.to_path(self.package_dir());
|
let patch_path = patch_path.to_path(self.package_dir());
|
||||||
|
|
||||||
|
@ -102,7 +101,6 @@ impl Project {
|
||||||
tracing::warn!(
|
tracing::warn!(
|
||||||
"patch for {name}@{version_id} not applied because it is not in the graph"
|
"patch for {name}@{version_id} not applied because it is not in the graph"
|
||||||
);
|
);
|
||||||
tx.send(Ok(format!("{name}@{version_id}"))).await.unwrap();
|
|
||||||
continue;
|
continue;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -115,41 +113,23 @@ impl Project {
|
||||||
version_id.version(),
|
version_id.version(),
|
||||||
);
|
);
|
||||||
|
|
||||||
tokio::spawn(async move {
|
let reporter = reporter.clone();
|
||||||
|
|
||||||
|
tasks.spawn(async move {
|
||||||
tracing::debug!("applying patch to {name}@{version_id}");
|
tracing::debug!("applying patch to {name}@{version_id}");
|
||||||
|
|
||||||
let patch = match fs::read(&patch_path).await {
|
let display_name = format!("{name}@{version_id}");
|
||||||
Ok(patch) => patch,
|
let progress_reporter = reporter.report_patch(&display_name);
|
||||||
Err(e) => {
|
|
||||||
tx.send(Err(errors::ApplyPatchesError::PatchRead(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let patch = match Diff::from_buffer(&patch) {
|
let patch = fs::read(&patch_path)
|
||||||
Ok(patch) => patch,
|
.await
|
||||||
Err(e) => {
|
.map_err(errors::ApplyPatchesError::PatchRead)?;
|
||||||
tx.send(Err(errors::ApplyPatchesError::Git(e)))
|
let patch = Diff::from_buffer(&patch)?;
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
{
|
{
|
||||||
let repo = match setup_patches_repo(&container_folder) {
|
let repo = setup_patches_repo(&container_folder)?;
|
||||||
Ok(repo) => repo,
|
|
||||||
Err(e) => {
|
|
||||||
tx.send(Err(errors::ApplyPatchesError::Git(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let modified_files = patch
|
let mut apply_delta_tasks = patch
|
||||||
.deltas()
|
.deltas()
|
||||||
.filter(|delta| matches!(delta.status(), git2::Delta::Modified))
|
.filter(|delta| matches!(delta.status(), git2::Delta::Modified))
|
||||||
.filter_map(|delta| delta.new_file().path())
|
.filter_map(|delta| delta.new_file().path())
|
||||||
|
@ -159,61 +139,45 @@ impl Project {
|
||||||
.to_path(&container_folder)
|
.to_path(&container_folder)
|
||||||
})
|
})
|
||||||
.filter(|path| path.is_file())
|
.filter(|path| path.is_file())
|
||||||
.collect::<Vec<_>>();
|
.map(|path| {
|
||||||
|
async {
|
||||||
for path in modified_files {
|
// so, we always unlink it
|
||||||
// there is no way (as far as I know) to check if it's hardlinked
|
let content = fs::read(&path).await?;
|
||||||
// so, we always unlink it
|
fs::remove_file(&path).await?;
|
||||||
let content = match fs::read(&path).await {
|
fs::write(path, content).await?;
|
||||||
Ok(content) => content,
|
Ok(())
|
||||||
Err(e) => {
|
|
||||||
tx.send(Err(errors::ApplyPatchesError::File(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
};
|
.map_err(errors::ApplyPatchesError::File)
|
||||||
|
})
|
||||||
|
.collect::<JoinSet<_>>();
|
||||||
|
|
||||||
if let Err(e) = fs::remove_file(&path).await {
|
while let Some(res) = apply_delta_tasks.join_next().await {
|
||||||
tx.send(Err(errors::ApplyPatchesError::File(e)))
|
res.unwrap()?;
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Err(e) = fs::write(path, content).await {
|
|
||||||
tx.send(Err(errors::ApplyPatchesError::File(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if let Err(e) = repo.apply(&patch, ApplyLocation::Both, None) {
|
repo.apply(&patch, ApplyLocation::Both, None)?;
|
||||||
tx.send(Err(errors::ApplyPatchesError::Git(e)))
|
|
||||||
.await
|
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
tracing::debug!(
|
tracing::debug!(
|
||||||
"patch applied to {name}@{version_id}, removing .git directory"
|
"patch applied to {name}@{version_id}, removing .git directory"
|
||||||
);
|
);
|
||||||
|
|
||||||
if let Err(e) = fs::remove_dir_all(container_folder.join(".git")).await {
|
fs::remove_dir_all(container_folder.join(".git"))
|
||||||
tx.send(Err(errors::ApplyPatchesError::DotGitRemove(e)))
|
.await
|
||||||
.await
|
.map_err(errors::ApplyPatchesError::DotGitRemove)?;
|
||||||
.unwrap();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
tx.send(Ok(format!("{name}@{version_id}"))).await.unwrap();
|
progress_reporter.report_done();
|
||||||
|
|
||||||
|
Ok(())
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(rx)
|
while let Some(res) = tasks.join_next().await {
|
||||||
|
res.unwrap()?
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
63
src/reporters.rs
Normal file
63
src/reporters.rs
Normal file
|
@ -0,0 +1,63 @@
|
||||||
|
//! Progress reporting
|
||||||
|
//!
|
||||||
|
//! Certain operations will ask for a progress reporter to be passed in, this
|
||||||
|
//! allows the caller to be notified of progress during the operation. This can
|
||||||
|
//! be used to show progress to the user.
|
||||||
|
//!
|
||||||
|
//! All reporter traits are implemented for `()`. These implementations do
|
||||||
|
//! nothing, and can be used to ignore progress reporting.
|
||||||
|
|
||||||
|
#![allow(unused_variables)]
|
||||||
|
|
||||||
|
/// Reports downloads.
|
||||||
|
pub trait DownloadsReporter<'a>: Send + Sync {
|
||||||
|
/// The [`DownloadProgressReporter`] type associated with this reporter.
|
||||||
|
type DownloadProgressReporter: DownloadProgressReporter + 'a;
|
||||||
|
|
||||||
|
/// Starts a new download.
|
||||||
|
fn report_download<'b>(&'a self, name: &'b str) -> Self::DownloadProgressReporter;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DownloadsReporter<'_> for () {
|
||||||
|
type DownloadProgressReporter = ();
|
||||||
|
fn report_download(&self, name: &str) -> Self::DownloadProgressReporter {}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Reports the progress of a single download.
|
||||||
|
pub trait DownloadProgressReporter: Send + Sync {
|
||||||
|
/// Reports that the download has started.
|
||||||
|
fn report_start(&self) {}
|
||||||
|
|
||||||
|
/// Reports the progress of the download.
|
||||||
|
///
|
||||||
|
/// `total` is the total number of bytes to download, and `len` is the number
|
||||||
|
/// of bytes downloaded so far.
|
||||||
|
fn report_progress(&self, total: u64, len: u64) {}
|
||||||
|
|
||||||
|
/// Reports that the download is done.
|
||||||
|
fn report_done(&self) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DownloadProgressReporter for () {}
|
||||||
|
|
||||||
|
/// Reports the progress of applying patches.
|
||||||
|
pub trait PatchesReporter<'a>: Send + Sync {
|
||||||
|
/// The [`PatchProgressReporter`] type associated with this reporter.
|
||||||
|
type PatchProgressReporter: PatchProgressReporter + 'a;
|
||||||
|
|
||||||
|
/// Starts a new patch.
|
||||||
|
fn report_patch<'b>(&'a self, name: &'b str) -> Self::PatchProgressReporter;
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PatchesReporter<'_> for () {
|
||||||
|
type PatchProgressReporter = ();
|
||||||
|
fn report_patch(&self, name: &str) -> Self::PatchProgressReporter {}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Reports the progress of a single patch.
|
||||||
|
pub trait PatchProgressReporter: Send + Sync {
|
||||||
|
/// Reports that the patch has been applied.
|
||||||
|
fn report_done(&self) {}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl PatchProgressReporter for () {}
|
|
@ -4,6 +4,7 @@ use crate::{
|
||||||
Manifest,
|
Manifest,
|
||||||
},
|
},
|
||||||
names::PackageNames,
|
names::PackageNames,
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{
|
source::{
|
||||||
fs::{store_in_cas, FSEntry, PackageFS},
|
fs::{store_in_cas, FSEntry, PackageFS},
|
||||||
git::{pkg_ref::GitPackageRef, specifier::GitDependencySpecifier},
|
git::{pkg_ref::GitPackageRef, specifier::GitDependencySpecifier},
|
||||||
|
@ -338,6 +339,7 @@ impl PackageSource for GitPackageSource {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
_reqwest: &reqwest::Client,
|
_reqwest: &reqwest::Client,
|
||||||
|
_reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
||||||
let index_file = project
|
let index_file = project
|
||||||
.cas_dir
|
.cas_dir
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
use crate::{
|
use crate::{
|
||||||
manifest::target::{Target, TargetKind},
|
manifest::target::{Target, TargetKind},
|
||||||
names::PackageNames,
|
names::PackageNames,
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{
|
source::{
|
||||||
fs::PackageFS, refs::PackageRefs, specifiers::DependencySpecifiers, traits::*,
|
fs::PackageFS, refs::PackageRefs, specifiers::DependencySpecifiers, traits::*,
|
||||||
version_id::VersionId,
|
version_id::VersionId,
|
||||||
|
@ -10,6 +11,7 @@ use crate::{
|
||||||
use std::{
|
use std::{
|
||||||
collections::{BTreeMap, HashSet},
|
collections::{BTreeMap, HashSet},
|
||||||
fmt::Debug,
|
fmt::Debug,
|
||||||
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
|
||||||
/// Packages' filesystems
|
/// Packages' filesystems
|
||||||
|
@ -152,26 +154,27 @@ impl PackageSource for PackageSources {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
reqwest: &reqwest::Client,
|
reqwest: &reqwest::Client,
|
||||||
|
reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
||||||
match (self, pkg_ref) {
|
match (self, pkg_ref) {
|
||||||
(PackageSources::Pesde(source), PackageRefs::Pesde(pkg_ref)) => source
|
(PackageSources::Pesde(source), PackageRefs::Pesde(pkg_ref)) => source
|
||||||
.download(pkg_ref, project, reqwest)
|
.download(pkg_ref, project, reqwest, reporter)
|
||||||
.await
|
.await
|
||||||
.map_err(Into::into),
|
.map_err(Into::into),
|
||||||
|
|
||||||
#[cfg(feature = "wally-compat")]
|
#[cfg(feature = "wally-compat")]
|
||||||
(PackageSources::Wally(source), PackageRefs::Wally(pkg_ref)) => source
|
(PackageSources::Wally(source), PackageRefs::Wally(pkg_ref)) => source
|
||||||
.download(pkg_ref, project, reqwest)
|
.download(pkg_ref, project, reqwest, reporter)
|
||||||
.await
|
.await
|
||||||
.map_err(Into::into),
|
.map_err(Into::into),
|
||||||
|
|
||||||
(PackageSources::Git(source), PackageRefs::Git(pkg_ref)) => source
|
(PackageSources::Git(source), PackageRefs::Git(pkg_ref)) => source
|
||||||
.download(pkg_ref, project, reqwest)
|
.download(pkg_ref, project, reqwest, reporter)
|
||||||
.await
|
.await
|
||||||
.map_err(Into::into),
|
.map_err(Into::into),
|
||||||
|
|
||||||
(PackageSources::Workspace(source), PackageRefs::Workspace(pkg_ref)) => source
|
(PackageSources::Workspace(source), PackageRefs::Workspace(pkg_ref)) => source
|
||||||
.download(pkg_ref, project, reqwest)
|
.download(pkg_ref, project, reqwest, reporter)
|
||||||
.await
|
.await
|
||||||
.map_err(Into::into),
|
.map_err(Into::into),
|
||||||
|
|
||||||
|
|
|
@ -7,7 +7,9 @@ use std::{
|
||||||
fmt::Debug,
|
fmt::Debug,
|
||||||
hash::Hash,
|
hash::Hash,
|
||||||
path::PathBuf,
|
path::PathBuf,
|
||||||
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
use tokio_util::io::StreamReader;
|
||||||
|
|
||||||
use pkg_ref::PesdePackageRef;
|
use pkg_ref::PesdePackageRef;
|
||||||
use specifier::PesdeDependencySpecifier;
|
use specifier::PesdeDependencySpecifier;
|
||||||
|
@ -18,6 +20,7 @@ use crate::{
|
||||||
DependencyType,
|
DependencyType,
|
||||||
},
|
},
|
||||||
names::{PackageName, PackageNames},
|
names::{PackageName, PackageNames},
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{
|
source::{
|
||||||
fs::{store_in_cas, FSEntry, PackageFS},
|
fs::{store_in_cas, FSEntry, PackageFS},
|
||||||
git_index::{read_file, root_tree, GitBasedSource},
|
git_index::{read_file, root_tree, GitBasedSource},
|
||||||
|
@ -165,6 +168,7 @@ impl PackageSource for PesdePackageSource {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
reqwest: &reqwest::Client,
|
reqwest: &reqwest::Client,
|
||||||
|
reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
||||||
let config = self.config(project).await.map_err(Box::new)?;
|
let config = self.config(project).await.map_err(Box::new)?;
|
||||||
let index_file = project
|
let index_file = project
|
||||||
|
@ -202,9 +206,26 @@ impl PackageSource for PesdePackageSource {
|
||||||
}
|
}
|
||||||
|
|
||||||
let response = request.send().await?.error_for_status()?;
|
let response = request.send().await?.error_for_status()?;
|
||||||
let bytes = response.bytes().await?;
|
|
||||||
|
|
||||||
let mut decoder = async_compression::tokio::bufread::GzipDecoder::new(bytes.as_ref());
|
let total_len = response.content_length().unwrap_or(0);
|
||||||
|
reporter.report_progress(total_len, 0);
|
||||||
|
|
||||||
|
let mut bytes_downloaded = 0;
|
||||||
|
let bytes = response
|
||||||
|
.bytes_stream()
|
||||||
|
.inspect(|chunk| {
|
||||||
|
chunk.as_ref().ok().inspect(|chunk| {
|
||||||
|
bytes_downloaded += chunk.len() as u64;
|
||||||
|
reporter.report_progress(total_len, bytes_downloaded);
|
||||||
|
});
|
||||||
|
})
|
||||||
|
.map(|result| {
|
||||||
|
result.map_err(|err| std::io::Error::new(std::io::ErrorKind::Other, err))
|
||||||
|
});
|
||||||
|
|
||||||
|
let bytes = StreamReader::new(bytes);
|
||||||
|
|
||||||
|
let mut decoder = async_compression::tokio::bufread::GzipDecoder::new(bytes);
|
||||||
let mut archive = tokio_tar::Archive::new(&mut decoder);
|
let mut archive = tokio_tar::Archive::new(&mut decoder);
|
||||||
|
|
||||||
let mut entries = BTreeMap::new();
|
let mut entries = BTreeMap::new();
|
||||||
|
@ -254,6 +275,8 @@ impl PackageSource for PesdePackageSource {
|
||||||
.await
|
.await
|
||||||
.map_err(errors::DownloadError::WriteIndex)?;
|
.map_err(errors::DownloadError::WriteIndex)?;
|
||||||
|
|
||||||
|
reporter.report_done();
|
||||||
|
|
||||||
Ok((fs, pkg_ref.target.clone()))
|
Ok((fs, pkg_ref.target.clone()))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -4,12 +4,14 @@ use crate::{
|
||||||
target::{Target, TargetKind},
|
target::{Target, TargetKind},
|
||||||
DependencyType,
|
DependencyType,
|
||||||
},
|
},
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{DependencySpecifiers, PackageFS, PackageSources, ResolveResult},
|
source::{DependencySpecifiers, PackageFS, PackageSources, ResolveResult},
|
||||||
Project,
|
Project,
|
||||||
};
|
};
|
||||||
use std::{
|
use std::{
|
||||||
collections::{BTreeMap, HashSet},
|
collections::{BTreeMap, HashSet},
|
||||||
fmt::{Debug, Display},
|
fmt::{Debug, Display},
|
||||||
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
|
||||||
/// A specifier for a dependency
|
/// A specifier for a dependency
|
||||||
|
@ -58,5 +60,6 @@ pub trait PackageSource: Debug {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
reqwest: &reqwest::Client,
|
reqwest: &reqwest::Client,
|
||||||
|
reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError>;
|
) -> Result<(PackageFS, Target), Self::DownloadError>;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
use crate::{
|
use crate::{
|
||||||
manifest::target::{Target, TargetKind},
|
manifest::target::{Target, TargetKind},
|
||||||
names::PackageNames,
|
names::PackageNames,
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{
|
source::{
|
||||||
fs::{store_in_cas, FSEntry, PackageFS},
|
fs::{store_in_cas, FSEntry, PackageFS},
|
||||||
git_index::{read_file, root_tree, GitBasedSource},
|
git_index::{read_file, root_tree, GitBasedSource},
|
||||||
|
@ -17,7 +18,7 @@ use crate::{
|
||||||
Project,
|
Project,
|
||||||
};
|
};
|
||||||
use fs_err::tokio as fs;
|
use fs_err::tokio as fs;
|
||||||
use futures::future::try_join_all;
|
use futures::{future::try_join_all, StreamExt};
|
||||||
use gix::Url;
|
use gix::Url;
|
||||||
use relative_path::RelativePathBuf;
|
use relative_path::RelativePathBuf;
|
||||||
use reqwest::header::AUTHORIZATION;
|
use reqwest::header::AUTHORIZATION;
|
||||||
|
@ -28,8 +29,12 @@ use std::{
|
||||||
sync::Arc,
|
sync::Arc,
|
||||||
};
|
};
|
||||||
use tempfile::tempdir;
|
use tempfile::tempdir;
|
||||||
use tokio::{io::AsyncWriteExt, sync::Mutex, task::spawn_blocking};
|
use tokio::{
|
||||||
use tokio_util::compat::FuturesAsyncReadCompatExt;
|
io::{AsyncReadExt, AsyncWriteExt},
|
||||||
|
sync::Mutex,
|
||||||
|
task::spawn_blocking,
|
||||||
|
};
|
||||||
|
use tokio_util::{compat::FuturesAsyncReadCompatExt, io::StreamReader};
|
||||||
use tracing::instrument;
|
use tracing::instrument;
|
||||||
|
|
||||||
pub(crate) mod compat_util;
|
pub(crate) mod compat_util;
|
||||||
|
@ -202,6 +207,7 @@ impl PackageSource for WallyPackageSource {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
reqwest: &reqwest::Client,
|
reqwest: &reqwest::Client,
|
||||||
|
reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
||||||
let config = self.config(project).await.map_err(Box::new)?;
|
let config = self.config(project).await.map_err(Box::new)?;
|
||||||
let index_file = project
|
let index_file = project
|
||||||
|
@ -250,12 +256,30 @@ impl PackageSource for WallyPackageSource {
|
||||||
}
|
}
|
||||||
|
|
||||||
let response = request.send().await?.error_for_status()?;
|
let response = request.send().await?.error_for_status()?;
|
||||||
let mut bytes = response.bytes().await?;
|
|
||||||
|
|
||||||
let archive = async_zip::tokio::read::seek::ZipFileReader::with_tokio(
|
let total_len = response.content_length().unwrap_or(0);
|
||||||
std::io::Cursor::new(&mut bytes),
|
reporter.report_progress(total_len, 0);
|
||||||
)
|
|
||||||
.await?;
|
let mut bytes_downloaded = 0;
|
||||||
|
let bytes = response
|
||||||
|
.bytes_stream()
|
||||||
|
.inspect(|chunk| {
|
||||||
|
chunk.as_ref().ok().inspect(|chunk| {
|
||||||
|
bytes_downloaded += chunk.len() as u64;
|
||||||
|
reporter.report_progress(total_len, bytes_downloaded);
|
||||||
|
});
|
||||||
|
})
|
||||||
|
.map(|result| {
|
||||||
|
result.map_err(|err| std::io::Error::new(std::io::ErrorKind::Other, err))
|
||||||
|
});
|
||||||
|
|
||||||
|
let mut bytes = StreamReader::new(bytes);
|
||||||
|
let mut buf = vec![];
|
||||||
|
bytes.read_to_end(&mut buf).await?;
|
||||||
|
|
||||||
|
let archive =
|
||||||
|
async_zip::tokio::read::seek::ZipFileReader::with_tokio(std::io::Cursor::new(&mut buf))
|
||||||
|
.await?;
|
||||||
|
|
||||||
let entries = (0..archive.file().entries().len())
|
let entries = (0..archive.file().entries().len())
|
||||||
.map(|index| {
|
.map(|index| {
|
||||||
|
@ -328,6 +352,8 @@ impl PackageSource for WallyPackageSource {
|
||||||
.await
|
.await
|
||||||
.map_err(errors::DownloadError::WriteIndex)?;
|
.map_err(errors::DownloadError::WriteIndex)?;
|
||||||
|
|
||||||
|
reporter.report_done();
|
||||||
|
|
||||||
Ok((fs, get_target(project, &tempdir).await?))
|
Ok((fs, get_target(project, &tempdir).await?))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
use crate::{
|
use crate::{
|
||||||
manifest::target::{Target, TargetKind},
|
manifest::target::{Target, TargetKind},
|
||||||
names::PackageNames,
|
names::PackageNames,
|
||||||
|
reporters::DownloadProgressReporter,
|
||||||
source::{
|
source::{
|
||||||
fs::PackageFS, specifiers::DependencySpecifiers, traits::PackageSource,
|
fs::PackageFS, specifiers::DependencySpecifiers, traits::PackageSource,
|
||||||
version_id::VersionId, workspace::pkg_ref::WorkspacePackageRef, PackageSources,
|
version_id::VersionId, workspace::pkg_ref::WorkspacePackageRef, PackageSources,
|
||||||
|
@ -11,7 +12,10 @@ use crate::{
|
||||||
use futures::StreamExt;
|
use futures::StreamExt;
|
||||||
use relative_path::RelativePathBuf;
|
use relative_path::RelativePathBuf;
|
||||||
use reqwest::Client;
|
use reqwest::Client;
|
||||||
use std::collections::{BTreeMap, HashSet};
|
use std::{
|
||||||
|
collections::{BTreeMap, HashSet},
|
||||||
|
sync::Arc,
|
||||||
|
};
|
||||||
use tokio::pin;
|
use tokio::pin;
|
||||||
use tracing::instrument;
|
use tracing::instrument;
|
||||||
|
|
||||||
|
@ -134,6 +138,7 @@ impl PackageSource for WorkspacePackageSource {
|
||||||
pkg_ref: &Self::Ref,
|
pkg_ref: &Self::Ref,
|
||||||
project: &Project,
|
project: &Project,
|
||||||
_reqwest: &Client,
|
_reqwest: &Client,
|
||||||
|
_reporter: Arc<impl DownloadProgressReporter>,
|
||||||
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
) -> Result<(PackageFS, Target), Self::DownloadError> {
|
||||||
let path = pkg_ref.path.to_path(project.workspace_dir.clone().unwrap());
|
let path = pkg_ref.path.to_path(project.workspace_dir.clone().unwrap());
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue