feat: improve linking process

This commit is contained in:
daimond113 2025-01-02 21:30:25 +01:00
parent e5b629e0c5
commit 6f5e2a2473
No known key found for this signature in database
GPG key ID: 3A8ECE51328B513C
10 changed files with 214 additions and 245 deletions

View file

@ -17,6 +17,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
### Changed ### Changed
- Change handling of graphs to a flat structure by @daimond113 - Change handling of graphs to a flat structure by @daimond113
- Store dependency over downloaded graphs in the lockfile by @daimond113 - Store dependency over downloaded graphs in the lockfile by @daimond113
- Improve linking process by @daimond113
### Removed ### Removed
- Remove old includes format compatibility by @daimond113 - Remove old includes format compatibility by @daimond113

View file

@ -203,8 +203,7 @@ impl ExecuteCommand {
DownloadAndLinkOptions::<CliReporter<Stderr>, ()>::new(reqwest) DownloadAndLinkOptions::<CliReporter<Stderr>, ()>::new(reqwest)
.reporter(reporter) .reporter(reporter)
.refreshed_sources(refreshed_sources) .refreshed_sources(refreshed_sources)
.prod(true) .prod(true),
.write(true),
) )
.await .await
.context("failed to download and link dependencies")?; .context("failed to download and link dependencies")?;

View file

@ -5,6 +5,7 @@ use std::{
time::Instant, time::Instant,
}; };
use super::files::make_executable;
use crate::cli::{ use crate::cli::{
bin_dir, bin_dir,
reporters::{self, CliReporter}, reporters::{self, CliReporter},
@ -15,15 +16,13 @@ use colored::Colorize;
use fs_err::tokio as fs; use fs_err::tokio as fs;
use pesde::{ use pesde::{
download_and_link::{DownloadAndLinkHooks, DownloadAndLinkOptions}, download_and_link::{DownloadAndLinkHooks, DownloadAndLinkOptions},
graph::{DependencyGraph, DownloadedGraph}, graph::{DependencyGraph, DependencyGraphWithTarget},
lockfile::Lockfile, lockfile::Lockfile,
manifest::{target::TargetKind, DependencyType}, manifest::{target::TargetKind, DependencyType},
Project, RefreshedSources, LOCKFILE_FILE_NAME, MANIFEST_FILE_NAME, Project, RefreshedSources, LOCKFILE_FILE_NAME, MANIFEST_FILE_NAME,
}; };
use tokio::task::JoinSet; use tokio::task::JoinSet;
use super::files::make_executable;
fn bin_link_file(alias: &str) -> String { fn bin_link_file(alias: &str) -> String {
let mut all_combinations = BTreeSet::new(); let mut all_combinations = BTreeSet::new();
@ -63,11 +62,11 @@ impl DownloadAndLinkHooks for InstallHooks {
async fn on_bins_downloaded( async fn on_bins_downloaded(
&self, &self,
downloaded_graph: &DownloadedGraph, graph: &DependencyGraphWithTarget,
) -> Result<(), Self::Error> { ) -> Result<(), Self::Error> {
let mut tasks = downloaded_graph let mut tasks = graph
.values() .values()
.filter(|node| node.target.as_ref().is_some_and(|t| t.bin_path().is_some())) .filter(|node| node.target.bin_path().is_some())
.filter_map(|node| node.node.direct.as_ref()) .filter_map(|node| node.node.direct.as_ref())
.map(|(alias, _, _)| alias) .map(|(alias, _, _)| alias)
.filter(|alias| { .filter(|alias| {
@ -249,51 +248,46 @@ pub async fn install(
.context("failed to build dependency graph")?; .context("failed to build dependency graph")?;
let graph = Arc::new(graph); let graph = Arc::new(graph);
root_progress.reset();
root_progress.set_length(0);
root_progress.set_message("download");
root_progress.set_style(reporters::root_progress_style_with_progress());
let hooks = InstallHooks {
bin_folder: bin_dir().await?,
};
#[allow(unused_variables)]
let downloaded_graph = project
.download_and_link(
&graph,
DownloadAndLinkOptions::<CliReporter, InstallHooks>::new(reqwest.clone())
.reporter(
#[cfg(feature = "patches")]
reporter.clone(),
#[cfg(not(feature = "patches"))]
reporter,
)
.hooks(hooks)
.refreshed_sources(refreshed_sources)
.prod(options.prod)
.write(options.write)
.network_concurrency(options.network_concurrency),
)
.await
.context("failed to download and link dependencies")?;
#[cfg(feature = "patches")]
if options.write { if options.write {
use pesde::{download_and_link::filter_graph, graph::ConvertableGraph};
root_progress.reset(); root_progress.reset();
root_progress.set_length(0); root_progress.set_length(0);
root_progress.set_message("patch"); root_progress.set_message("download");
root_progress.set_style(reporters::root_progress_style_with_progress());
project let hooks = InstallHooks {
.apply_patches( bin_folder: bin_dir().await?,
&Arc::into_inner(filter_graph(&downloaded_graph, options.prod)) };
.unwrap()
.convert(), #[allow(unused_variables)]
reporter, let downloaded_graph = project
.download_and_link(
&graph,
DownloadAndLinkOptions::<CliReporter, InstallHooks>::new(reqwest.clone())
.reporter(
#[cfg(feature = "patches")]
reporter.clone(),
#[cfg(not(feature = "patches"))]
reporter,
)
.hooks(hooks)
.refreshed_sources(refreshed_sources)
.prod(options.prod)
.network_concurrency(options.network_concurrency),
) )
.await?; .await
.context("failed to download and link dependencies")?;
#[cfg(feature = "patches")]
{
use pesde::graph::ConvertableGraph;
root_progress.reset();
root_progress.set_length(0);
root_progress.set_message("patch");
project
.apply_patches(&downloaded_graph.convert(), reporter)
.await?;
}
} }
root_progress.set_message("finish"); root_progress.set_message("finish");

View file

@ -1,15 +1,14 @@
use crate::{ use crate::{
graph::{DependencyGraph, DownloadedDependencyGraphNode}, graph::{DependencyGraph, DependencyGraphNode},
manifest::DependencyType,
reporters::{DownloadProgressReporter, DownloadsReporter}, reporters::{DownloadProgressReporter, DownloadsReporter},
source::{ source::{
fs::PackageFs,
ids::PackageId, ids::PackageId,
traits::{DownloadOptions, GetTargetOptions, PackageRef, PackageSource, RefreshOptions}, traits::{DownloadOptions, PackageRef, PackageSource, RefreshOptions},
}, },
Project, RefreshedSources, PACKAGES_CONTAINER_NAME, Project, RefreshedSources,
}; };
use async_stream::try_stream; use async_stream::try_stream;
use fs_err::tokio as fs;
use futures::Stream; use futures::Stream;
use std::{num::NonZeroUsize, sync::Arc}; use std::{num::NonZeroUsize, sync::Arc};
use tokio::{sync::Semaphore, task::JoinSet}; use tokio::{sync::Semaphore, task::JoinSet};
@ -24,12 +23,6 @@ pub(crate) struct DownloadGraphOptions<Reporter> {
pub reporter: Option<Arc<Reporter>>, pub reporter: Option<Arc<Reporter>>,
/// The refreshed sources. /// The refreshed sources.
pub refreshed_sources: RefreshedSources, pub refreshed_sources: RefreshedSources,
/// Whether to skip dev dependencies.
pub prod: bool,
/// Whether to write the downloaded packages to disk.
pub write: bool,
/// Whether to download Wally packages.
pub wally: bool,
/// The max number of concurrent network requests. /// The max number of concurrent network requests.
pub network_concurrency: NonZeroUsize, pub network_concurrency: NonZeroUsize,
} }
@ -44,9 +37,6 @@ where
reqwest, reqwest,
reporter: None, reporter: None,
refreshed_sources: Default::default(), refreshed_sources: Default::default(),
prod: false,
write: false,
wally: false,
network_concurrency: NonZeroUsize::new(16).unwrap(), network_concurrency: NonZeroUsize::new(16).unwrap(),
} }
} }
@ -63,24 +53,6 @@ where
self self
} }
/// Sets whether to skip dev dependencies.
pub(crate) fn prod(mut self, prod: bool) -> Self {
self.prod = prod;
self
}
/// Sets whether to write the downloaded packages to disk.
pub(crate) fn write(mut self, write: bool) -> Self {
self.write = write;
self
}
/// Sets whether to download Wally packages.
pub(crate) fn wally(mut self, wally: bool) -> Self {
self.wally = wally;
self
}
/// Sets the max number of concurrent network requests. /// Sets the max number of concurrent network requests.
pub(crate) fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self { pub(crate) fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self {
self.network_concurrency = network_concurrency; self.network_concurrency = network_concurrency;
@ -94,9 +66,6 @@ impl<Reporter> Clone for DownloadGraphOptions<Reporter> {
reqwest: self.reqwest.clone(), reqwest: self.reqwest.clone(),
reporter: self.reporter.clone(), reporter: self.reporter.clone(),
refreshed_sources: self.refreshed_sources.clone(), refreshed_sources: self.refreshed_sources.clone(),
prod: self.prod,
write: self.write,
wally: self.wally,
network_concurrency: self.network_concurrency, network_concurrency: self.network_concurrency,
} }
} }
@ -104,14 +73,14 @@ impl<Reporter> Clone for DownloadGraphOptions<Reporter> {
impl Project { impl Project {
/// Downloads a graph of dependencies. /// Downloads a graph of dependencies.
#[instrument(skip_all, fields(prod = options.prod, wally = options.wally, write = options.write), level = "debug")] #[instrument(skip_all, level = "debug")]
pub(crate) async fn download_graph<Reporter>( pub(crate) async fn download_graph<Reporter>(
&self, &self,
graph: &DependencyGraph, graph: &DependencyGraph,
options: DownloadGraphOptions<Reporter>, options: DownloadGraphOptions<Reporter>,
) -> Result< ) -> Result<
impl Stream< impl Stream<
Item = Result<(DownloadedDependencyGraphNode, PackageId), errors::DownloadGraphError>, Item = Result<(PackageId, DependencyGraphNode, PackageFs), errors::DownloadGraphError>,
>, >,
errors::DownloadGraphError, errors::DownloadGraphError,
> >
@ -122,21 +91,13 @@ impl Project {
reqwest, reqwest,
reporter, reporter,
refreshed_sources, refreshed_sources,
prod,
write,
wally,
network_concurrency, network_concurrency,
} = options; } = options;
let manifest = self.deser_manifest().await?;
let manifest_target_kind = manifest.target.kind();
let semaphore = Arc::new(Semaphore::new(network_concurrency.get())); let semaphore = Arc::new(Semaphore::new(network_concurrency.get()));
let mut tasks = graph let mut tasks = graph
.iter() .iter()
// we need to download pesde packages first, since scripts (for target finding for example) can depend on them
.filter(|(_, node)| node.pkg_ref.like_wally() == wally)
.map(|(package_id, node)| { .map(|(package_id, node)| {
let span = tracing::info_span!("download", package_id = package_id.to_string()); let span = tracing::info_span!("download", package_id = package_id.to_string());
@ -144,7 +105,6 @@ impl Project {
let reqwest = reqwest.clone(); let reqwest = reqwest.clone();
let reporter = reporter.clone(); let reporter = reporter.clone();
let refreshed_sources = refreshed_sources.clone(); let refreshed_sources = refreshed_sources.clone();
let package_dir = project.package_dir().to_path_buf();
let semaphore = semaphore.clone(); let semaphore = semaphore.clone();
let package_id = Arc::new(package_id.clone()); let package_id = Arc::new(package_id.clone());
let node = node.clone(); let node = node.clone();
@ -170,15 +130,6 @@ impl Project {
) )
.await?; .await?;
let container_folder = package_dir
.join(
manifest_target_kind.packages_folder(package_id.version_id().target()),
)
.join(PACKAGES_CONTAINER_NAME)
.join(node.container_folder(&package_id));
fs::create_dir_all(&container_folder).await?;
tracing::debug!("downloading"); tracing::debug!("downloading");
let fs = match progress_reporter { let fs = match progress_reporter {
@ -213,33 +164,7 @@ impl Project {
tracing::debug!("downloaded"); tracing::debug!("downloaded");
let mut target = None; Ok((Arc::into_inner(package_id).unwrap(), node, fs))
if write {
if !prod || node.resolved_ty != DependencyType::Dev {
fs.write_to(&container_folder, project.cas_dir(), true)
.await?;
target = Some(
source
.get_target(
&node.pkg_ref,
&GetTargetOptions {
project,
path: Arc::from(container_folder),
id: package_id.clone(),
},
)
.await
.map_err(Box::new)?,
);
} else {
tracing::debug!("skipping write to disk, dev dependency in prod mode");
}
}
let downloaded_node = DownloadedDependencyGraphNode { node, target };
Ok((downloaded_node, Arc::into_inner(package_id).unwrap()))
} }
.instrument(span) .instrument(span)
}) })
@ -263,10 +188,6 @@ pub mod errors {
#[derive(Debug, Error)] #[derive(Debug, Error)]
#[non_exhaustive] #[non_exhaustive]
pub enum DownloadGraphError { pub enum DownloadGraphError {
/// An error occurred deserializing the project manifest
#[error("error deserializing project manifest")]
ManifestDeserializationFailed(#[from] crate::errors::ManifestReadError),
/// An error occurred refreshing a package source /// An error occurred refreshing a package source
#[error("failed to refresh package source")] #[error("failed to refresh package source")]
RefreshFailed(#[from] crate::source::errors::RefreshError), RefreshFailed(#[from] crate::source::errors::RefreshError),
@ -278,9 +199,5 @@ pub mod errors {
/// Error downloading a package /// Error downloading a package
#[error("failed to download package")] #[error("failed to download package")]
DownloadFailed(#[from] Box<crate::source::errors::DownloadError>), DownloadFailed(#[from] Box<crate::source::errors::DownloadError>),
/// Error getting target
#[error("failed to get target")]
GetTargetFailed(#[from] Box<crate::source::errors::GetTargetError>),
} }
} }

View file

@ -1,34 +1,30 @@
use crate::{ use crate::{
download::DownloadGraphOptions, download::DownloadGraphOptions,
graph::{DependencyGraph, DownloadedGraph}, graph::{
DependencyGraph, DependencyGraphNode, DependencyGraphNodeWithTarget,
DependencyGraphWithTarget,
},
manifest::DependencyType, manifest::DependencyType,
reporters::DownloadsReporter, reporters::DownloadsReporter,
Project, RefreshedSources, source::{
ids::PackageId,
traits::{GetTargetOptions, PackageRef, PackageSource},
},
Project, RefreshedSources, PACKAGES_CONTAINER_NAME,
}; };
use fs_err::tokio as fs;
use futures::TryStreamExt; use futures::TryStreamExt;
use std::{ use std::{
collections::{BTreeMap, HashMap},
convert::Infallible, convert::Infallible,
future::{self, Future}, future::{self, Future},
num::NonZeroUsize, num::NonZeroUsize,
path::PathBuf,
sync::Arc, sync::Arc,
}; };
use tokio::{pin, task::JoinSet};
use tracing::{instrument, Instrument}; use tracing::{instrument, Instrument};
/// Filters a graph to only include production dependencies, if `prod` is `true`
pub fn filter_graph(graph: &DownloadedGraph, prod: bool) -> Arc<DownloadedGraph> {
if !prod {
return Arc::new(graph.clone());
}
Arc::new(
graph
.iter()
.filter(|(_, node)| node.node.resolved_ty != DependencyType::Dev)
.map(|(id, node)| (id.clone(), node.clone()))
.collect(),
)
}
/// Hooks to perform actions after certain events during download and linking. /// Hooks to perform actions after certain events during download and linking.
#[allow(unused_variables)] #[allow(unused_variables)]
pub trait DownloadAndLinkHooks { pub trait DownloadAndLinkHooks {
@ -39,7 +35,7 @@ pub trait DownloadAndLinkHooks {
/// contains all downloaded packages. /// contains all downloaded packages.
fn on_scripts_downloaded( fn on_scripts_downloaded(
&self, &self,
downloaded_graph: &DownloadedGraph, graph: &DependencyGraphWithTarget,
) -> impl Future<Output = Result<(), Self::Error>> + Send { ) -> impl Future<Output = Result<(), Self::Error>> + Send {
future::ready(Ok(())) future::ready(Ok(()))
} }
@ -48,7 +44,7 @@ pub trait DownloadAndLinkHooks {
/// `downloaded_graph` contains all downloaded packages. /// `downloaded_graph` contains all downloaded packages.
fn on_bins_downloaded( fn on_bins_downloaded(
&self, &self,
downloaded_graph: &DownloadedGraph, graph: &DependencyGraphWithTarget,
) -> impl Future<Output = Result<(), Self::Error>> + Send { ) -> impl Future<Output = Result<(), Self::Error>> + Send {
future::ready(Ok(())) future::ready(Ok(()))
} }
@ -57,7 +53,7 @@ pub trait DownloadAndLinkHooks {
/// `downloaded_graph` contains all downloaded packages. /// `downloaded_graph` contains all downloaded packages.
fn on_all_downloaded( fn on_all_downloaded(
&self, &self,
downloaded_graph: &DownloadedGraph, graph: &DependencyGraphWithTarget,
) -> impl Future<Output = Result<(), Self::Error>> + Send { ) -> impl Future<Output = Result<(), Self::Error>> + Send {
future::ready(Ok(())) future::ready(Ok(()))
} }
@ -80,8 +76,6 @@ pub struct DownloadAndLinkOptions<Reporter = (), Hooks = ()> {
pub refreshed_sources: RefreshedSources, pub refreshed_sources: RefreshedSources,
/// Whether to skip dev dependencies. /// Whether to skip dev dependencies.
pub prod: bool, pub prod: bool,
/// Whether to write the downloaded packages to disk.
pub write: bool,
/// The max number of concurrent network requests. /// The max number of concurrent network requests.
pub network_concurrency: NonZeroUsize, pub network_concurrency: NonZeroUsize,
} }
@ -99,7 +93,6 @@ where
hooks: None, hooks: None,
refreshed_sources: Default::default(), refreshed_sources: Default::default(),
prod: false, prod: false,
write: true,
network_concurrency: NonZeroUsize::new(16).unwrap(), network_concurrency: NonZeroUsize::new(16).unwrap(),
} }
} }
@ -128,12 +121,6 @@ where
self self
} }
/// Sets whether to write the downloaded packages to disk.
pub fn write(mut self, write: bool) -> Self {
self.write = write;
self
}
/// Sets the max number of concurrent network requests. /// Sets the max number of concurrent network requests.
pub fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self { pub fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self {
self.network_concurrency = network_concurrency; self.network_concurrency = network_concurrency;
@ -149,7 +136,6 @@ impl Clone for DownloadAndLinkOptions {
hooks: self.hooks.clone(), hooks: self.hooks.clone(),
refreshed_sources: self.refreshed_sources.clone(), refreshed_sources: self.refreshed_sources.clone(),
prod: self.prod, prod: self.prod,
write: self.write,
network_concurrency: self.network_concurrency, network_concurrency: self.network_concurrency,
} }
} }
@ -157,12 +143,12 @@ impl Clone for DownloadAndLinkOptions {
impl Project { impl Project {
/// Downloads a graph of dependencies and links them in the correct order /// Downloads a graph of dependencies and links them in the correct order
#[instrument(skip_all, fields(prod = options.prod, write = options.write), level = "debug")] #[instrument(skip_all, fields(prod = options.prod), level = "debug")]
pub async fn download_and_link<Reporter, Hooks>( pub async fn download_and_link<Reporter, Hooks>(
&self, &self,
graph: &Arc<DependencyGraph>, graph: &Arc<DependencyGraph>,
options: DownloadAndLinkOptions<Reporter, Hooks>, options: DownloadAndLinkOptions<Reporter, Hooks>,
) -> Result<DownloadedGraph, errors::DownloadAndLinkError<Hooks::Error>> ) -> Result<DependencyGraphWithTarget, errors::DownloadAndLinkError<Hooks::Error>>
where where
Reporter: for<'a> DownloadsReporter<'a> + 'static, Reporter: for<'a> DownloadsReporter<'a> + 'static,
Hooks: DownloadAndLinkHooks + 'static, Hooks: DownloadAndLinkHooks + 'static,
@ -173,81 +159,151 @@ impl Project {
hooks, hooks,
refreshed_sources, refreshed_sources,
prod, prod,
write,
network_concurrency, network_concurrency,
} = options; } = options;
let graph = graph.clone(); let graph = graph.clone();
let reqwest = reqwest.clone(); let reqwest = reqwest.clone();
let manifest = self.deser_manifest().await?;
let mut downloaded_graph = DownloadedGraph::new(); // step 1. download dependencies
let downloaded_graph = {
let mut downloaded_graph = BTreeMap::new();
let mut download_graph_options = DownloadGraphOptions::<Reporter>::new(reqwest.clone()) let mut download_graph_options = DownloadGraphOptions::<Reporter>::new(reqwest.clone())
.refreshed_sources(refreshed_sources.clone()) .refreshed_sources(refreshed_sources.clone())
.prod(prod) .network_concurrency(network_concurrency);
.write(write)
.network_concurrency(network_concurrency);
if let Some(reporter) = reporter { if let Some(reporter) = reporter {
download_graph_options = download_graph_options.reporter(reporter.clone()); download_graph_options = download_graph_options.reporter(reporter.clone());
}
let downloaded = self
.download_graph(&graph, download_graph_options.clone())
.instrument(tracing::debug_span!("download"))
.await?;
pin!(downloaded);
let mut tasks = JoinSet::new();
while let Some((id, node, fs)) = downloaded.try_next().await? {
let container_folder = self
.package_dir()
.join(
manifest
.target
.kind()
.packages_folder(id.version_id().target()),
)
.join(PACKAGES_CONTAINER_NAME)
.join(node.container_folder(&id));
if prod && node.resolved_ty == DependencyType::Dev {
continue;
}
downloaded_graph.insert(id, (node, container_folder.clone()));
let cas_dir = self.cas_dir().to_path_buf();
tasks.spawn(async move {
fs::create_dir_all(&container_folder).await?;
fs.write_to(container_folder, cas_dir, true).await
});
}
while let Some(task) = tasks.join_next().await {
task.unwrap()?;
}
downloaded_graph
};
let (downloaded_wally_graph, downloaded_other_graph) = downloaded_graph
.into_iter()
.partition::<HashMap<_, _>, _>(|(_, (node, _))| node.pkg_ref.is_wally_package());
let mut graph = Arc::new(DependencyGraphWithTarget::new());
async fn get_graph_targets<Hooks: DownloadAndLinkHooks>(
graph: &mut Arc<DependencyGraphWithTarget>,
project: &Project,
downloaded_graph: HashMap<PackageId, (DependencyGraphNode, PathBuf)>,
) -> Result<(), errors::DownloadAndLinkError<Hooks::Error>> {
let mut tasks = downloaded_graph
.into_iter()
.map(|(id, (node, container_folder))| {
let source = node.pkg_ref.source();
let path = Arc::from(container_folder.as_path());
let id = Arc::new(id.clone());
let project = project.clone();
async move {
let target = source
.get_target(
&node.pkg_ref,
&GetTargetOptions {
project,
path,
id: id.clone(),
},
)
.await?;
Ok::<_, errors::DownloadAndLinkError<Hooks::Error>>((
Arc::into_inner(id).unwrap(),
DependencyGraphNodeWithTarget { node, target },
))
}
})
.collect::<JoinSet<_>>();
while let Some(task) = tasks.join_next().await {
let (id, node) = task.unwrap()?;
Arc::get_mut(graph).unwrap().insert(id, node);
}
Ok(())
} }
// step 1. download pesde dependencies // step 2. get targets for non Wally packages (Wally packages require the scripts packages to be downloaded first)
self.download_graph(&graph, download_graph_options.clone()) get_graph_targets::<Hooks>(&mut graph, self, downloaded_other_graph)
.instrument(tracing::debug_span!("download (pesde)")) .instrument(tracing::debug_span!("get targets (non-wally)"))
.await?
.try_for_each(|(downloaded_node, id)| {
downloaded_graph.insert(id, downloaded_node);
future::ready(Ok(()))
})
.await?; .await?;
// step 2. link pesde dependencies. do so without types self.link_dependencies(graph.clone(), false)
if write { .instrument(tracing::debug_span!("link (non-wally)"))
self.link_dependencies(filter_graph(&downloaded_graph, prod), false) .await?;
.instrument(tracing::debug_span!("link (pesde)"))
.await?;
}
if let Some(ref hooks) = hooks { if let Some(hooks) = &hooks {
hooks hooks
.on_scripts_downloaded(&downloaded_graph) .on_scripts_downloaded(&graph)
.await .await
.map_err(errors::DownloadAndLinkError::Hook)?; .map_err(errors::DownloadAndLinkError::Hook)?;
hooks hooks
.on_bins_downloaded(&downloaded_graph) .on_bins_downloaded(&graph)
.await .await
.map_err(errors::DownloadAndLinkError::Hook)?; .map_err(errors::DownloadAndLinkError::Hook)?;
} }
// step 3. download wally dependencies // step 3. get targets for Wally packages
self.download_graph(&graph, download_graph_options.clone().wally(true)) get_graph_targets::<Hooks>(&mut graph, self, downloaded_wally_graph)
.instrument(tracing::debug_span!("download (wally)")) .instrument(tracing::debug_span!("get targets (wally)"))
.await?
.try_for_each(|(downloaded_node, id)| {
downloaded_graph.insert(id, downloaded_node);
future::ready(Ok(()))
})
.await?; .await?;
// step 4. link ALL dependencies. do so with types // step 4. link ALL dependencies. do so with types
if write { self.link_dependencies(graph.clone(), true)
self.link_dependencies(filter_graph(&downloaded_graph, prod), true) .instrument(tracing::debug_span!("link (all)"))
.instrument(tracing::debug_span!("link (all)")) .await?;
.await?;
}
if let Some(ref hooks) = hooks { if let Some(hooks) = &hooks {
hooks hooks
.on_all_downloaded(&downloaded_graph) .on_all_downloaded(&graph)
.await .await
.map_err(errors::DownloadAndLinkError::Hook)?; .map_err(errors::DownloadAndLinkError::Hook)?;
} }
Ok(downloaded_graph) Ok(Arc::into_inner(graph).unwrap())
} }
} }
@ -259,6 +315,10 @@ pub mod errors {
#[derive(Debug, Error)] #[derive(Debug, Error)]
#[non_exhaustive] #[non_exhaustive]
pub enum DownloadAndLinkError<E> { pub enum DownloadAndLinkError<E> {
/// Reading the manifest failed
#[error("error reading manifest")]
ManifestRead(#[from] crate::errors::ManifestReadError),
/// An error occurred while downloading the graph /// An error occurred while downloading the graph
#[error("error downloading graph")] #[error("error downloading graph")]
DownloadGraph(#[from] crate::download::errors::DownloadGraphError), DownloadGraph(#[from] crate::download::errors::DownloadGraphError),
@ -270,5 +330,13 @@ pub mod errors {
/// An error occurred while executing the pesde callback /// An error occurred while executing the pesde callback
#[error("error executing hook")] #[error("error executing hook")]
Hook(#[source] E), Hook(#[source] E),
/// IO error
#[error("io error")]
Io(#[from] std::io::Error),
/// Error getting a target
#[error("error getting target")]
GetTarget(#[from] crate::source::errors::GetTargetError),
} }
} }

View file

@ -47,7 +47,7 @@ impl DependencyGraphNode {
pub fn container_folder(&self, package_id: &PackageId) -> PathBuf { pub fn container_folder(&self, package_id: &PackageId) -> PathBuf {
let (name, version) = package_id.parts(); let (name, version) = package_id.parts();
if self.pkg_ref.like_wally() { if self.pkg_ref.is_wally_package() {
return PathBuf::from(format!( return PathBuf::from(format!(
"{}_{}@{}", "{}_{}@{}",
name.as_str().0, name.as_str().0,
@ -66,18 +66,17 @@ impl DependencyGraphNode {
/// A graph of `DependencyGraphNode`s /// A graph of `DependencyGraphNode`s
pub type DependencyGraph = Graph<DependencyGraphNode>; pub type DependencyGraph = Graph<DependencyGraphNode>;
/// A downloaded dependency graph node, i.e. a `DependencyGraphNode` with a `Target` /// A dependency graph node with a `Target`
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
pub struct DownloadedDependencyGraphNode { pub struct DependencyGraphNodeWithTarget {
/// The target of the package /// The target of the package
/// None only if download was called with write = false or is a dev dependency in a prod install pub target: Target,
pub target: Option<Target>,
/// The node /// The node
pub node: DependencyGraphNode, pub node: DependencyGraphNode,
} }
/// A graph of `DownloadedDependencyGraphNode`s /// A graph of `DownloadedDependencyGraphNode`s
pub type DownloadedGraph = Graph<DownloadedDependencyGraphNode>; pub type DependencyGraphWithTarget = Graph<DependencyGraphNodeWithTarget>;
/// A trait for converting a graph to a different type of graph /// A trait for converting a graph to a different type of graph
pub trait ConvertableGraph<Node> { pub trait ConvertableGraph<Node> {
@ -85,7 +84,7 @@ pub trait ConvertableGraph<Node> {
fn convert(self) -> Graph<Node>; fn convert(self) -> Graph<Node>;
} }
impl ConvertableGraph<DependencyGraphNode> for DownloadedGraph { impl ConvertableGraph<DependencyGraphNode> for DependencyGraphWithTarget {
fn convert(self) -> Graph<DependencyGraphNode> { fn convert(self) -> Graph<DependencyGraphNode> {
self.into_iter().map(|(id, node)| (id, node.node)).collect() self.into_iter().map(|(id, node)| (id, node.node)).collect()
} }

View file

@ -118,7 +118,7 @@ fn luau_style_path(path: &Path) -> String {
/// Get the require path for a library /// Get the require path for a library
#[allow(clippy::too_many_arguments)] #[allow(clippy::too_many_arguments)]
pub fn get_lib_require_path( pub fn get_lib_require_path(
target: &TargetKind, target: TargetKind,
base_dir: &Path, base_dir: &Path,
lib_file: &RelativePath, lib_file: &RelativePath,
destination_dir: &Path, destination_dir: &Path,

View file

@ -1,5 +1,5 @@
use crate::{ use crate::{
graph::{DownloadedDependencyGraphNode, DownloadedGraph}, graph::{DependencyGraphNodeWithTarget, DependencyGraphWithTarget},
linking::generator::get_file_types, linking::generator::get_file_types,
manifest::Manifest, manifest::Manifest,
scripts::{execute_script, ExecuteScriptHooks, ScriptName}, scripts::{execute_script, ExecuteScriptHooks, ScriptName},
@ -59,7 +59,7 @@ impl Project {
#[instrument(skip(self, graph), level = "debug")] #[instrument(skip(self, graph), level = "debug")]
pub(crate) async fn link_dependencies( pub(crate) async fn link_dependencies(
&self, &self,
graph: Arc<DownloadedGraph>, graph: Arc<DependencyGraphWithTarget>,
with_types: bool, with_types: bool,
) -> Result<(), errors::LinkingError> { ) -> Result<(), errors::LinkingError> {
let manifest = self.deser_manifest().await?; let manifest = self.deser_manifest().await?;
@ -87,7 +87,7 @@ impl Project {
let project = self.clone(); let project = self.clone();
async move { async move {
let Some(lib_file) = node.target.as_ref().and_then(|t| t.lib_path()) else { let Some(lib_file) = node.target.lib_path() else {
return Ok((package_id, vec![])); return Ok((package_id, vec![]));
}; };
@ -123,10 +123,8 @@ impl Project {
vec![] vec![]
}; };
if let Some(build_files) = node if let Some(build_files) = Some(&node.target)
.target .filter(|_| !node.node.pkg_ref.is_wally_package())
.as_ref()
.filter(|_| !node.node.pkg_ref.like_wally())
.and_then(|t| t.build_files()) .and_then(|t| t.build_files())
{ {
execute_script( execute_script(
@ -165,7 +163,7 @@ impl Project {
container_folder: &Path, container_folder: &Path,
root_container_folder: &Path, root_container_folder: &Path,
relative_container_folder: &Path, relative_container_folder: &Path,
node: &DownloadedDependencyGraphNode, node: &DependencyGraphNodeWithTarget,
package_id: &PackageId, package_id: &PackageId,
alias: &str, alias: &str,
package_types: &PackageTypes, package_types: &PackageTypes,
@ -173,14 +171,10 @@ impl Project {
) -> Result<(), errors::LinkingError> { ) -> Result<(), errors::LinkingError> {
static NO_TYPES: Vec<String> = Vec::new(); static NO_TYPES: Vec<String> = Vec::new();
let Some(target) = &node.target else { if let Some(lib_file) = node.target.lib_path() {
return Ok(());
};
if let Some(lib_file) = target.lib_path() {
let lib_module = generator::generate_lib_linking_module( let lib_module = generator::generate_lib_linking_module(
&generator::get_lib_require_path( &generator::get_lib_require_path(
&target.kind(), node.target.kind(),
base_folder, base_folder,
lib_file, lib_file,
container_folder, container_folder,
@ -200,7 +194,7 @@ impl Project {
.await?; .await?;
} }
if let Some(bin_file) = target.bin_path() { if let Some(bin_file) = node.target.bin_path() {
let bin_module = generator::generate_bin_linking_module( let bin_module = generator::generate_bin_linking_module(
container_folder, container_folder,
&generator::get_bin_require_path(base_folder, bin_file, container_folder), &generator::get_bin_require_path(base_folder, bin_file, container_folder),
@ -214,7 +208,7 @@ impl Project {
.await?; .await?;
} }
if let Some(scripts) = target.scripts().filter(|s| !s.is_empty()) { if let Some(scripts) = node.target.scripts().filter(|s| !s.is_empty()) {
let scripts_base = let scripts_base =
create_and_canonicalize(self.package_dir().join(SCRIPTS_LINK_FOLDER).join(alias)) create_and_canonicalize(self.package_dir().join(SCRIPTS_LINK_FOLDER).join(alias))
.await?; .await?;
@ -241,7 +235,7 @@ impl Project {
async fn link( async fn link(
&self, &self,
graph: &Arc<DownloadedGraph>, graph: &Arc<DependencyGraphWithTarget>,
manifest: &Arc<Manifest>, manifest: &Arc<Manifest>,
package_types: &Arc<PackageTypes>, package_types: &Arc<PackageTypes>,
is_complete: bool, is_complete: bool,
@ -322,10 +316,7 @@ impl Project {
let linker_folder = create_and_canonicalize(node_container_folder.join( let linker_folder = create_and_canonicalize(node_container_folder.join(
node.node.base_folder( node.node.base_folder(
package_id.version_id(), package_id.version_id(),
match &dependency_node.target { dependency_node.target.kind(),
Some(t) => t.kind(),
None => continue,
},
), ),
)) ))
.await?; .await?;

View file

@ -209,7 +209,7 @@ pub enum RobloxPlaceKind {
Server, Server,
} }
impl TryInto<RobloxPlaceKind> for &TargetKind { impl TryInto<RobloxPlaceKind> for TargetKind {
type Error = (); type Error = ();
fn try_into(self) -> Result<RobloxPlaceKind, Self::Error> { fn try_into(self) -> Result<RobloxPlaceKind, Self::Error> {

View file

@ -24,7 +24,7 @@ pub enum PackageRefs {
impl PackageRefs { impl PackageRefs {
/// Returns whether this package reference should be treated as a Wally package /// Returns whether this package reference should be treated as a Wally package
pub fn like_wally(&self) -> bool { pub fn is_wally_package(&self) -> bool {
match self { match self {
#[cfg(feature = "wally-compat")] #[cfg(feature = "wally-compat")]
PackageRefs::Wally(_) => true, PackageRefs::Wally(_) => true,