From 6f5e2a247394c84da2c5fa894f03009df9e1169b Mon Sep 17 00:00:00 2001 From: daimond113 <72147841+daimond113@users.noreply.github.com> Date: Thu, 2 Jan 2025 21:30:25 +0100 Subject: [PATCH] feat: improve linking process --- CHANGELOG.md | 1 + src/cli/commands/execute.rs | 3 +- src/cli/install.rs | 86 +++++++------- src/download.rs | 97 ++-------------- src/download_and_link.rs | 220 +++++++++++++++++++++++------------- src/graph.rs | 13 +-- src/linking/generator.rs | 2 +- src/linking/mod.rs | 33 ++---- src/manifest/target.rs | 2 +- src/source/refs.rs | 2 +- 10 files changed, 214 insertions(+), 245 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5059197..d7301ab 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -17,6 +17,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 ### Changed - Change handling of graphs to a flat structure by @daimond113 - Store dependency over downloaded graphs in the lockfile by @daimond113 +- Improve linking process by @daimond113 ### Removed - Remove old includes format compatibility by @daimond113 diff --git a/src/cli/commands/execute.rs b/src/cli/commands/execute.rs index 0bcaf69..3cbe2fc 100644 --- a/src/cli/commands/execute.rs +++ b/src/cli/commands/execute.rs @@ -203,8 +203,7 @@ impl ExecuteCommand { DownloadAndLinkOptions::, ()>::new(reqwest) .reporter(reporter) .refreshed_sources(refreshed_sources) - .prod(true) - .write(true), + .prod(true), ) .await .context("failed to download and link dependencies")?; diff --git a/src/cli/install.rs b/src/cli/install.rs index 1c04ddb..0649795 100644 --- a/src/cli/install.rs +++ b/src/cli/install.rs @@ -5,6 +5,7 @@ use std::{ time::Instant, }; +use super::files::make_executable; use crate::cli::{ bin_dir, reporters::{self, CliReporter}, @@ -15,15 +16,13 @@ use colored::Colorize; use fs_err::tokio as fs; use pesde::{ download_and_link::{DownloadAndLinkHooks, DownloadAndLinkOptions}, - graph::{DependencyGraph, DownloadedGraph}, + graph::{DependencyGraph, DependencyGraphWithTarget}, lockfile::Lockfile, manifest::{target::TargetKind, DependencyType}, Project, RefreshedSources, LOCKFILE_FILE_NAME, MANIFEST_FILE_NAME, }; use tokio::task::JoinSet; -use super::files::make_executable; - fn bin_link_file(alias: &str) -> String { let mut all_combinations = BTreeSet::new(); @@ -63,11 +62,11 @@ impl DownloadAndLinkHooks for InstallHooks { async fn on_bins_downloaded( &self, - downloaded_graph: &DownloadedGraph, + graph: &DependencyGraphWithTarget, ) -> Result<(), Self::Error> { - let mut tasks = downloaded_graph + let mut tasks = graph .values() - .filter(|node| node.target.as_ref().is_some_and(|t| t.bin_path().is_some())) + .filter(|node| node.target.bin_path().is_some()) .filter_map(|node| node.node.direct.as_ref()) .map(|(alias, _, _)| alias) .filter(|alias| { @@ -249,51 +248,46 @@ pub async fn install( .context("failed to build dependency graph")?; let graph = Arc::new(graph); - root_progress.reset(); - root_progress.set_length(0); - root_progress.set_message("download"); - root_progress.set_style(reporters::root_progress_style_with_progress()); - - let hooks = InstallHooks { - bin_folder: bin_dir().await?, - }; - - #[allow(unused_variables)] - let downloaded_graph = project - .download_and_link( - &graph, - DownloadAndLinkOptions::::new(reqwest.clone()) - .reporter( - #[cfg(feature = "patches")] - reporter.clone(), - #[cfg(not(feature = "patches"))] - reporter, - ) - .hooks(hooks) - .refreshed_sources(refreshed_sources) - .prod(options.prod) - .write(options.write) - .network_concurrency(options.network_concurrency), - ) - .await - .context("failed to download and link dependencies")?; - - #[cfg(feature = "patches")] if options.write { - use pesde::{download_and_link::filter_graph, graph::ConvertableGraph}; - root_progress.reset(); root_progress.set_length(0); - root_progress.set_message("patch"); + root_progress.set_message("download"); + root_progress.set_style(reporters::root_progress_style_with_progress()); - project - .apply_patches( - &Arc::into_inner(filter_graph(&downloaded_graph, options.prod)) - .unwrap() - .convert(), - reporter, + let hooks = InstallHooks { + bin_folder: bin_dir().await?, + }; + + #[allow(unused_variables)] + let downloaded_graph = project + .download_and_link( + &graph, + DownloadAndLinkOptions::::new(reqwest.clone()) + .reporter( + #[cfg(feature = "patches")] + reporter.clone(), + #[cfg(not(feature = "patches"))] + reporter, + ) + .hooks(hooks) + .refreshed_sources(refreshed_sources) + .prod(options.prod) + .network_concurrency(options.network_concurrency), ) - .await?; + .await + .context("failed to download and link dependencies")?; + + #[cfg(feature = "patches")] + { + use pesde::graph::ConvertableGraph; + root_progress.reset(); + root_progress.set_length(0); + root_progress.set_message("patch"); + + project + .apply_patches(&downloaded_graph.convert(), reporter) + .await?; + } } root_progress.set_message("finish"); diff --git a/src/download.rs b/src/download.rs index db1bf54..6b5e6c2 100644 --- a/src/download.rs +++ b/src/download.rs @@ -1,15 +1,14 @@ use crate::{ - graph::{DependencyGraph, DownloadedDependencyGraphNode}, - manifest::DependencyType, + graph::{DependencyGraph, DependencyGraphNode}, reporters::{DownloadProgressReporter, DownloadsReporter}, source::{ + fs::PackageFs, ids::PackageId, - traits::{DownloadOptions, GetTargetOptions, PackageRef, PackageSource, RefreshOptions}, + traits::{DownloadOptions, PackageRef, PackageSource, RefreshOptions}, }, - Project, RefreshedSources, PACKAGES_CONTAINER_NAME, + Project, RefreshedSources, }; use async_stream::try_stream; -use fs_err::tokio as fs; use futures::Stream; use std::{num::NonZeroUsize, sync::Arc}; use tokio::{sync::Semaphore, task::JoinSet}; @@ -24,12 +23,6 @@ pub(crate) struct DownloadGraphOptions { pub reporter: Option>, /// The refreshed sources. pub refreshed_sources: RefreshedSources, - /// Whether to skip dev dependencies. - pub prod: bool, - /// Whether to write the downloaded packages to disk. - pub write: bool, - /// Whether to download Wally packages. - pub wally: bool, /// The max number of concurrent network requests. pub network_concurrency: NonZeroUsize, } @@ -44,9 +37,6 @@ where reqwest, reporter: None, refreshed_sources: Default::default(), - prod: false, - write: false, - wally: false, network_concurrency: NonZeroUsize::new(16).unwrap(), } } @@ -63,24 +53,6 @@ where self } - /// Sets whether to skip dev dependencies. - pub(crate) fn prod(mut self, prod: bool) -> Self { - self.prod = prod; - self - } - - /// Sets whether to write the downloaded packages to disk. - pub(crate) fn write(mut self, write: bool) -> Self { - self.write = write; - self - } - - /// Sets whether to download Wally packages. - pub(crate) fn wally(mut self, wally: bool) -> Self { - self.wally = wally; - self - } - /// Sets the max number of concurrent network requests. pub(crate) fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self { self.network_concurrency = network_concurrency; @@ -94,9 +66,6 @@ impl Clone for DownloadGraphOptions { reqwest: self.reqwest.clone(), reporter: self.reporter.clone(), refreshed_sources: self.refreshed_sources.clone(), - prod: self.prod, - write: self.write, - wally: self.wally, network_concurrency: self.network_concurrency, } } @@ -104,14 +73,14 @@ impl Clone for DownloadGraphOptions { impl Project { /// Downloads a graph of dependencies. - #[instrument(skip_all, fields(prod = options.prod, wally = options.wally, write = options.write), level = "debug")] + #[instrument(skip_all, level = "debug")] pub(crate) async fn download_graph( &self, graph: &DependencyGraph, options: DownloadGraphOptions, ) -> Result< impl Stream< - Item = Result<(DownloadedDependencyGraphNode, PackageId), errors::DownloadGraphError>, + Item = Result<(PackageId, DependencyGraphNode, PackageFs), errors::DownloadGraphError>, >, errors::DownloadGraphError, > @@ -122,21 +91,13 @@ impl Project { reqwest, reporter, refreshed_sources, - prod, - write, - wally, network_concurrency, } = options; - let manifest = self.deser_manifest().await?; - let manifest_target_kind = manifest.target.kind(); - let semaphore = Arc::new(Semaphore::new(network_concurrency.get())); let mut tasks = graph .iter() - // we need to download pesde packages first, since scripts (for target finding for example) can depend on them - .filter(|(_, node)| node.pkg_ref.like_wally() == wally) .map(|(package_id, node)| { let span = tracing::info_span!("download", package_id = package_id.to_string()); @@ -144,7 +105,6 @@ impl Project { let reqwest = reqwest.clone(); let reporter = reporter.clone(); let refreshed_sources = refreshed_sources.clone(); - let package_dir = project.package_dir().to_path_buf(); let semaphore = semaphore.clone(); let package_id = Arc::new(package_id.clone()); let node = node.clone(); @@ -170,15 +130,6 @@ impl Project { ) .await?; - let container_folder = package_dir - .join( - manifest_target_kind.packages_folder(package_id.version_id().target()), - ) - .join(PACKAGES_CONTAINER_NAME) - .join(node.container_folder(&package_id)); - - fs::create_dir_all(&container_folder).await?; - tracing::debug!("downloading"); let fs = match progress_reporter { @@ -213,33 +164,7 @@ impl Project { tracing::debug!("downloaded"); - let mut target = None; - - if write { - if !prod || node.resolved_ty != DependencyType::Dev { - fs.write_to(&container_folder, project.cas_dir(), true) - .await?; - - target = Some( - source - .get_target( - &node.pkg_ref, - &GetTargetOptions { - project, - path: Arc::from(container_folder), - id: package_id.clone(), - }, - ) - .await - .map_err(Box::new)?, - ); - } else { - tracing::debug!("skipping write to disk, dev dependency in prod mode"); - } - } - - let downloaded_node = DownloadedDependencyGraphNode { node, target }; - Ok((downloaded_node, Arc::into_inner(package_id).unwrap())) + Ok((Arc::into_inner(package_id).unwrap(), node, fs)) } .instrument(span) }) @@ -263,10 +188,6 @@ pub mod errors { #[derive(Debug, Error)] #[non_exhaustive] pub enum DownloadGraphError { - /// An error occurred deserializing the project manifest - #[error("error deserializing project manifest")] - ManifestDeserializationFailed(#[from] crate::errors::ManifestReadError), - /// An error occurred refreshing a package source #[error("failed to refresh package source")] RefreshFailed(#[from] crate::source::errors::RefreshError), @@ -278,9 +199,5 @@ pub mod errors { /// Error downloading a package #[error("failed to download package")] DownloadFailed(#[from] Box), - - /// Error getting target - #[error("failed to get target")] - GetTargetFailed(#[from] Box), } } diff --git a/src/download_and_link.rs b/src/download_and_link.rs index 1b36ff4..9b1b37f 100644 --- a/src/download_and_link.rs +++ b/src/download_and_link.rs @@ -1,34 +1,30 @@ use crate::{ download::DownloadGraphOptions, - graph::{DependencyGraph, DownloadedGraph}, + graph::{ + DependencyGraph, DependencyGraphNode, DependencyGraphNodeWithTarget, + DependencyGraphWithTarget, + }, manifest::DependencyType, reporters::DownloadsReporter, - Project, RefreshedSources, + source::{ + ids::PackageId, + traits::{GetTargetOptions, PackageRef, PackageSource}, + }, + Project, RefreshedSources, PACKAGES_CONTAINER_NAME, }; +use fs_err::tokio as fs; use futures::TryStreamExt; use std::{ + collections::{BTreeMap, HashMap}, convert::Infallible, future::{self, Future}, num::NonZeroUsize, + path::PathBuf, sync::Arc, }; +use tokio::{pin, task::JoinSet}; use tracing::{instrument, Instrument}; -/// Filters a graph to only include production dependencies, if `prod` is `true` -pub fn filter_graph(graph: &DownloadedGraph, prod: bool) -> Arc { - if !prod { - return Arc::new(graph.clone()); - } - - Arc::new( - graph - .iter() - .filter(|(_, node)| node.node.resolved_ty != DependencyType::Dev) - .map(|(id, node)| (id.clone(), node.clone())) - .collect(), - ) -} - /// Hooks to perform actions after certain events during download and linking. #[allow(unused_variables)] pub trait DownloadAndLinkHooks { @@ -39,7 +35,7 @@ pub trait DownloadAndLinkHooks { /// contains all downloaded packages. fn on_scripts_downloaded( &self, - downloaded_graph: &DownloadedGraph, + graph: &DependencyGraphWithTarget, ) -> impl Future> + Send { future::ready(Ok(())) } @@ -48,7 +44,7 @@ pub trait DownloadAndLinkHooks { /// `downloaded_graph` contains all downloaded packages. fn on_bins_downloaded( &self, - downloaded_graph: &DownloadedGraph, + graph: &DependencyGraphWithTarget, ) -> impl Future> + Send { future::ready(Ok(())) } @@ -57,7 +53,7 @@ pub trait DownloadAndLinkHooks { /// `downloaded_graph` contains all downloaded packages. fn on_all_downloaded( &self, - downloaded_graph: &DownloadedGraph, + graph: &DependencyGraphWithTarget, ) -> impl Future> + Send { future::ready(Ok(())) } @@ -80,8 +76,6 @@ pub struct DownloadAndLinkOptions { pub refreshed_sources: RefreshedSources, /// Whether to skip dev dependencies. pub prod: bool, - /// Whether to write the downloaded packages to disk. - pub write: bool, /// The max number of concurrent network requests. pub network_concurrency: NonZeroUsize, } @@ -99,7 +93,6 @@ where hooks: None, refreshed_sources: Default::default(), prod: false, - write: true, network_concurrency: NonZeroUsize::new(16).unwrap(), } } @@ -128,12 +121,6 @@ where self } - /// Sets whether to write the downloaded packages to disk. - pub fn write(mut self, write: bool) -> Self { - self.write = write; - self - } - /// Sets the max number of concurrent network requests. pub fn network_concurrency(mut self, network_concurrency: NonZeroUsize) -> Self { self.network_concurrency = network_concurrency; @@ -149,7 +136,6 @@ impl Clone for DownloadAndLinkOptions { hooks: self.hooks.clone(), refreshed_sources: self.refreshed_sources.clone(), prod: self.prod, - write: self.write, network_concurrency: self.network_concurrency, } } @@ -157,12 +143,12 @@ impl Clone for DownloadAndLinkOptions { impl Project { /// Downloads a graph of dependencies and links them in the correct order - #[instrument(skip_all, fields(prod = options.prod, write = options.write), level = "debug")] + #[instrument(skip_all, fields(prod = options.prod), level = "debug")] pub async fn download_and_link( &self, graph: &Arc, options: DownloadAndLinkOptions, - ) -> Result> + ) -> Result> where Reporter: for<'a> DownloadsReporter<'a> + 'static, Hooks: DownloadAndLinkHooks + 'static, @@ -173,81 +159,151 @@ impl Project { hooks, refreshed_sources, prod, - write, network_concurrency, } = options; let graph = graph.clone(); let reqwest = reqwest.clone(); + let manifest = self.deser_manifest().await?; - let mut downloaded_graph = DownloadedGraph::new(); + // step 1. download dependencies + let downloaded_graph = { + let mut downloaded_graph = BTreeMap::new(); - let mut download_graph_options = DownloadGraphOptions::::new(reqwest.clone()) - .refreshed_sources(refreshed_sources.clone()) - .prod(prod) - .write(write) - .network_concurrency(network_concurrency); + let mut download_graph_options = DownloadGraphOptions::::new(reqwest.clone()) + .refreshed_sources(refreshed_sources.clone()) + .network_concurrency(network_concurrency); - if let Some(reporter) = reporter { - download_graph_options = download_graph_options.reporter(reporter.clone()); + if let Some(reporter) = reporter { + download_graph_options = download_graph_options.reporter(reporter.clone()); + } + + let downloaded = self + .download_graph(&graph, download_graph_options.clone()) + .instrument(tracing::debug_span!("download")) + .await?; + pin!(downloaded); + + let mut tasks = JoinSet::new(); + + while let Some((id, node, fs)) = downloaded.try_next().await? { + let container_folder = self + .package_dir() + .join( + manifest + .target + .kind() + .packages_folder(id.version_id().target()), + ) + .join(PACKAGES_CONTAINER_NAME) + .join(node.container_folder(&id)); + + if prod && node.resolved_ty == DependencyType::Dev { + continue; + } + + downloaded_graph.insert(id, (node, container_folder.clone())); + + let cas_dir = self.cas_dir().to_path_buf(); + tasks.spawn(async move { + fs::create_dir_all(&container_folder).await?; + fs.write_to(container_folder, cas_dir, true).await + }); + } + + while let Some(task) = tasks.join_next().await { + task.unwrap()?; + } + + downloaded_graph + }; + + let (downloaded_wally_graph, downloaded_other_graph) = downloaded_graph + .into_iter() + .partition::, _>(|(_, (node, _))| node.pkg_ref.is_wally_package()); + + let mut graph = Arc::new(DependencyGraphWithTarget::new()); + + async fn get_graph_targets( + graph: &mut Arc, + project: &Project, + downloaded_graph: HashMap, + ) -> Result<(), errors::DownloadAndLinkError> { + let mut tasks = downloaded_graph + .into_iter() + .map(|(id, (node, container_folder))| { + let source = node.pkg_ref.source(); + let path = Arc::from(container_folder.as_path()); + let id = Arc::new(id.clone()); + let project = project.clone(); + + async move { + let target = source + .get_target( + &node.pkg_ref, + &GetTargetOptions { + project, + path, + id: id.clone(), + }, + ) + .await?; + + Ok::<_, errors::DownloadAndLinkError>(( + Arc::into_inner(id).unwrap(), + DependencyGraphNodeWithTarget { node, target }, + )) + } + }) + .collect::>(); + + while let Some(task) = tasks.join_next().await { + let (id, node) = task.unwrap()?; + Arc::get_mut(graph).unwrap().insert(id, node); + } + + Ok(()) } - // step 1. download pesde dependencies - self.download_graph(&graph, download_graph_options.clone()) - .instrument(tracing::debug_span!("download (pesde)")) - .await? - .try_for_each(|(downloaded_node, id)| { - downloaded_graph.insert(id, downloaded_node); - - future::ready(Ok(())) - }) + // step 2. get targets for non Wally packages (Wally packages require the scripts packages to be downloaded first) + get_graph_targets::(&mut graph, self, downloaded_other_graph) + .instrument(tracing::debug_span!("get targets (non-wally)")) .await?; - // step 2. link pesde dependencies. do so without types - if write { - self.link_dependencies(filter_graph(&downloaded_graph, prod), false) - .instrument(tracing::debug_span!("link (pesde)")) - .await?; - } + self.link_dependencies(graph.clone(), false) + .instrument(tracing::debug_span!("link (non-wally)")) + .await?; - if let Some(ref hooks) = hooks { + if let Some(hooks) = &hooks { hooks - .on_scripts_downloaded(&downloaded_graph) + .on_scripts_downloaded(&graph) .await .map_err(errors::DownloadAndLinkError::Hook)?; hooks - .on_bins_downloaded(&downloaded_graph) + .on_bins_downloaded(&graph) .await .map_err(errors::DownloadAndLinkError::Hook)?; } - // step 3. download wally dependencies - self.download_graph(&graph, download_graph_options.clone().wally(true)) - .instrument(tracing::debug_span!("download (wally)")) - .await? - .try_for_each(|(downloaded_node, id)| { - downloaded_graph.insert(id, downloaded_node); - - future::ready(Ok(())) - }) + // step 3. get targets for Wally packages + get_graph_targets::(&mut graph, self, downloaded_wally_graph) + .instrument(tracing::debug_span!("get targets (wally)")) .await?; // step 4. link ALL dependencies. do so with types - if write { - self.link_dependencies(filter_graph(&downloaded_graph, prod), true) - .instrument(tracing::debug_span!("link (all)")) - .await?; - } + self.link_dependencies(graph.clone(), true) + .instrument(tracing::debug_span!("link (all)")) + .await?; - if let Some(ref hooks) = hooks { + if let Some(hooks) = &hooks { hooks - .on_all_downloaded(&downloaded_graph) + .on_all_downloaded(&graph) .await .map_err(errors::DownloadAndLinkError::Hook)?; } - Ok(downloaded_graph) + Ok(Arc::into_inner(graph).unwrap()) } } @@ -259,6 +315,10 @@ pub mod errors { #[derive(Debug, Error)] #[non_exhaustive] pub enum DownloadAndLinkError { + /// Reading the manifest failed + #[error("error reading manifest")] + ManifestRead(#[from] crate::errors::ManifestReadError), + /// An error occurred while downloading the graph #[error("error downloading graph")] DownloadGraph(#[from] crate::download::errors::DownloadGraphError), @@ -270,5 +330,13 @@ pub mod errors { /// An error occurred while executing the pesde callback #[error("error executing hook")] Hook(#[source] E), + + /// IO error + #[error("io error")] + Io(#[from] std::io::Error), + + /// Error getting a target + #[error("error getting target")] + GetTarget(#[from] crate::source::errors::GetTargetError), } } diff --git a/src/graph.rs b/src/graph.rs index 82cc70e..bf04746 100644 --- a/src/graph.rs +++ b/src/graph.rs @@ -47,7 +47,7 @@ impl DependencyGraphNode { pub fn container_folder(&self, package_id: &PackageId) -> PathBuf { let (name, version) = package_id.parts(); - if self.pkg_ref.like_wally() { + if self.pkg_ref.is_wally_package() { return PathBuf::from(format!( "{}_{}@{}", name.as_str().0, @@ -66,18 +66,17 @@ impl DependencyGraphNode { /// A graph of `DependencyGraphNode`s pub type DependencyGraph = Graph; -/// A downloaded dependency graph node, i.e. a `DependencyGraphNode` with a `Target` +/// A dependency graph node with a `Target` #[derive(Debug, Clone)] -pub struct DownloadedDependencyGraphNode { +pub struct DependencyGraphNodeWithTarget { /// The target of the package - /// None only if download was called with write = false or is a dev dependency in a prod install - pub target: Option, + pub target: Target, /// The node pub node: DependencyGraphNode, } /// A graph of `DownloadedDependencyGraphNode`s -pub type DownloadedGraph = Graph; +pub type DependencyGraphWithTarget = Graph; /// A trait for converting a graph to a different type of graph pub trait ConvertableGraph { @@ -85,7 +84,7 @@ pub trait ConvertableGraph { fn convert(self) -> Graph; } -impl ConvertableGraph for DownloadedGraph { +impl ConvertableGraph for DependencyGraphWithTarget { fn convert(self) -> Graph { self.into_iter().map(|(id, node)| (id, node.node)).collect() } diff --git a/src/linking/generator.rs b/src/linking/generator.rs index ab54162..00c5d7c 100644 --- a/src/linking/generator.rs +++ b/src/linking/generator.rs @@ -118,7 +118,7 @@ fn luau_style_path(path: &Path) -> String { /// Get the require path for a library #[allow(clippy::too_many_arguments)] pub fn get_lib_require_path( - target: &TargetKind, + target: TargetKind, base_dir: &Path, lib_file: &RelativePath, destination_dir: &Path, diff --git a/src/linking/mod.rs b/src/linking/mod.rs index 0b3062a..37dd204 100644 --- a/src/linking/mod.rs +++ b/src/linking/mod.rs @@ -1,5 +1,5 @@ use crate::{ - graph::{DownloadedDependencyGraphNode, DownloadedGraph}, + graph::{DependencyGraphNodeWithTarget, DependencyGraphWithTarget}, linking::generator::get_file_types, manifest::Manifest, scripts::{execute_script, ExecuteScriptHooks, ScriptName}, @@ -59,7 +59,7 @@ impl Project { #[instrument(skip(self, graph), level = "debug")] pub(crate) async fn link_dependencies( &self, - graph: Arc, + graph: Arc, with_types: bool, ) -> Result<(), errors::LinkingError> { let manifest = self.deser_manifest().await?; @@ -87,7 +87,7 @@ impl Project { let project = self.clone(); async move { - let Some(lib_file) = node.target.as_ref().and_then(|t| t.lib_path()) else { + let Some(lib_file) = node.target.lib_path() else { return Ok((package_id, vec![])); }; @@ -123,10 +123,8 @@ impl Project { vec![] }; - if let Some(build_files) = node - .target - .as_ref() - .filter(|_| !node.node.pkg_ref.like_wally()) + if let Some(build_files) = Some(&node.target) + .filter(|_| !node.node.pkg_ref.is_wally_package()) .and_then(|t| t.build_files()) { execute_script( @@ -165,7 +163,7 @@ impl Project { container_folder: &Path, root_container_folder: &Path, relative_container_folder: &Path, - node: &DownloadedDependencyGraphNode, + node: &DependencyGraphNodeWithTarget, package_id: &PackageId, alias: &str, package_types: &PackageTypes, @@ -173,14 +171,10 @@ impl Project { ) -> Result<(), errors::LinkingError> { static NO_TYPES: Vec = Vec::new(); - let Some(target) = &node.target else { - return Ok(()); - }; - - if let Some(lib_file) = target.lib_path() { + if let Some(lib_file) = node.target.lib_path() { let lib_module = generator::generate_lib_linking_module( &generator::get_lib_require_path( - &target.kind(), + node.target.kind(), base_folder, lib_file, container_folder, @@ -200,7 +194,7 @@ impl Project { .await?; } - if let Some(bin_file) = target.bin_path() { + if let Some(bin_file) = node.target.bin_path() { let bin_module = generator::generate_bin_linking_module( container_folder, &generator::get_bin_require_path(base_folder, bin_file, container_folder), @@ -214,7 +208,7 @@ impl Project { .await?; } - if let Some(scripts) = target.scripts().filter(|s| !s.is_empty()) { + if let Some(scripts) = node.target.scripts().filter(|s| !s.is_empty()) { let scripts_base = create_and_canonicalize(self.package_dir().join(SCRIPTS_LINK_FOLDER).join(alias)) .await?; @@ -241,7 +235,7 @@ impl Project { async fn link( &self, - graph: &Arc, + graph: &Arc, manifest: &Arc, package_types: &Arc, is_complete: bool, @@ -322,10 +316,7 @@ impl Project { let linker_folder = create_and_canonicalize(node_container_folder.join( node.node.base_folder( package_id.version_id(), - match &dependency_node.target { - Some(t) => t.kind(), - None => continue, - }, + dependency_node.target.kind(), ), )) .await?; diff --git a/src/manifest/target.rs b/src/manifest/target.rs index a64b71f..1195600 100644 --- a/src/manifest/target.rs +++ b/src/manifest/target.rs @@ -209,7 +209,7 @@ pub enum RobloxPlaceKind { Server, } -impl TryInto for &TargetKind { +impl TryInto for TargetKind { type Error = (); fn try_into(self) -> Result { diff --git a/src/source/refs.rs b/src/source/refs.rs index 8b268c9..b6ebbb8 100644 --- a/src/source/refs.rs +++ b/src/source/refs.rs @@ -24,7 +24,7 @@ pub enum PackageRefs { impl PackageRefs { /// Returns whether this package reference should be treated as a Wally package - pub fn like_wally(&self) -> bool { + pub fn is_wally_package(&self) -> bool { match self { #[cfg(feature = "wally-compat")] PackageRefs::Wally(_) => true,