| use std::{ |
| borrow::Cow, |
| cmp::max, |
| collections::BinaryHeap, |
| fmt::{Debug, Display}, |
| future::Future, |
| mem::take, |
| path::{Path, PathBuf}, |
| process::{ExitStatus, Stdio}, |
| sync::Arc, |
| time::{Duration, Instant}, |
| }; |
|
|
| use anyhow::{Context, Result, bail}; |
| use futures::join; |
| use once_cell::sync::Lazy; |
| use owo_colors::{OwoColorize, Style}; |
| use parking_lot::Mutex; |
| use rustc_hash::FxHashMap; |
| use serde::{Serialize, de::DeserializeOwned}; |
| use tokio::{ |
| io::{ |
| AsyncBufReadExt, AsyncRead, AsyncReadExt, AsyncWrite, AsyncWriteExt, BufReader, Stderr, |
| Stdout, stderr, stdout, |
| }, |
| net::{TcpListener, TcpStream}, |
| process::{Child, ChildStderr, ChildStdout, Command}, |
| select, |
| sync::{OwnedSemaphorePermit, Semaphore}, |
| time::{sleep, timeout}, |
| }; |
| use turbo_rcstr::RcStr; |
| use turbo_tasks::{FxIndexSet, ResolvedVc, Vc, duration_span}; |
| use turbo_tasks_fs::{FileSystemPath, json::parse_json_with_source_context}; |
| use turbopack_ecmascript::magic_identifier::unmangle_identifiers; |
|
|
| use crate::{AssetsForSourceMapping, source_map::apply_source_mapping}; |
|
|
| #[derive(Clone, Copy)] |
| pub enum FormattingMode { |
| |
| Plain, |
| |
| AnsiColors, |
| } |
|
|
| impl FormattingMode { |
| pub fn magic_identifier<'a>(&self, content: impl Display + 'a) -> impl Display + 'a { |
| match self { |
| FormattingMode::Plain => format!("{{{content}}}"), |
| FormattingMode::AnsiColors => format!("{{{content}}}").italic().to_string(), |
| } |
| } |
|
|
| pub fn lowlight<'a>(&self, content: impl Display + 'a) -> impl Display + 'a { |
| match self { |
| FormattingMode::Plain => Style::new(), |
| FormattingMode::AnsiColors => Style::new().dimmed(), |
| } |
| .style(content) |
| } |
|
|
| pub fn highlight<'a>(&self, content: impl Display + 'a) -> impl Display + 'a { |
| match self { |
| FormattingMode::Plain => Style::new(), |
| FormattingMode::AnsiColors => Style::new().bold().underline(), |
| } |
| .style(content) |
| } |
| } |
|
|
| struct NodeJsPoolProcess { |
| child: Option<Child>, |
| connection: TcpStream, |
| assets_for_source_mapping: ResolvedVc<AssetsForSourceMapping>, |
| assets_root: FileSystemPath, |
| project_dir: FileSystemPath, |
| stdout_handler: OutputStreamHandler<ChildStdout, Stdout>, |
| stderr_handler: OutputStreamHandler<ChildStderr, Stderr>, |
| debug: bool, |
| cpu_time_invested: Duration, |
| } |
|
|
| impl Ord for NodeJsPoolProcess { |
| fn cmp(&self, other: &Self) -> std::cmp::Ordering { |
| self.cpu_time_invested |
| .cmp(&other.cpu_time_invested) |
| .then_with(|| { |
| self.child |
| .as_ref() |
| .map(|c| c.id()) |
| .cmp(&other.child.as_ref().map(|c| c.id())) |
| }) |
| } |
| } |
|
|
| impl PartialOrd for NodeJsPoolProcess { |
| fn partial_cmp(&self, other: &Self) -> Option<std::cmp::Ordering> { |
| Some(self.cmp(other)) |
| } |
| } |
|
|
| impl Eq for NodeJsPoolProcess {} |
|
|
| impl PartialEq for NodeJsPoolProcess { |
| fn eq(&self, other: &Self) -> bool { |
| self.cmp(other) == std::cmp::Ordering::Equal |
| } |
| } |
|
|
| impl NodeJsPoolProcess { |
| pub async fn apply_source_mapping<'a>( |
| &self, |
| text: &'a str, |
| formatting_mode: FormattingMode, |
| ) -> Result<Cow<'a, str>> { |
| let text = unmangle_identifiers(text, |content| formatting_mode.magic_identifier(content)); |
| match text { |
| Cow::Borrowed(text) => { |
| apply_source_mapping( |
| text, |
| *self.assets_for_source_mapping, |
| self.assets_root.clone(), |
| self.project_dir.clone(), |
| formatting_mode, |
| ) |
| .await |
| } |
| Cow::Owned(ref text) => { |
| let cow = apply_source_mapping( |
| text, |
| *self.assets_for_source_mapping, |
| self.assets_root.clone(), |
| self.project_dir.clone(), |
| formatting_mode, |
| ) |
| .await?; |
| Ok(Cow::Owned(cow.into_owned())) |
| } |
| } |
| } |
| } |
|
|
| const CONNECT_TIMEOUT: Duration = Duration::from_secs(30); |
|
|
| #[derive(Clone, PartialEq, Eq, Hash)] |
| struct OutputEntry { |
| data: Arc<[u8]>, |
| stack_trace: Option<Arc<[u8]>>, |
| } |
|
|
| type SharedOutputSet = Arc<Mutex<FxIndexSet<(OutputEntry, u32)>>>; |
|
|
| static GLOBAL_OUTPUT_LOCK: tokio::sync::Mutex<()> = tokio::sync::Mutex::const_new(()); |
| static MARKER: &[u8] = b"TURBOPACK_OUTPUT_"; |
| static MARKER_STR: &str = "TURBOPACK_OUTPUT_"; |
|
|
| struct OutputStreamHandler<R: AsyncRead + Unpin, W: AsyncWrite + Unpin> { |
| stream: BufReader<R>, |
| shared: SharedOutputSet, |
| assets_for_source_mapping: ResolvedVc<AssetsForSourceMapping>, |
| root: FileSystemPath, |
| project_dir: FileSystemPath, |
| final_stream: W, |
| } |
|
|
| impl<R: AsyncRead + Unpin, W: AsyncWrite + Unpin> OutputStreamHandler<R, W> { |
| |
| |
| |
| |
| pub async fn handle_operation(&mut self) -> Result<()> { |
| let Self { |
| stream, |
| shared, |
| assets_for_source_mapping, |
| root, |
| project_dir, |
| final_stream, |
| } = self; |
|
|
| async fn write_final<W: AsyncWrite + Unpin>( |
| mut bytes: &[u8], |
| final_stream: &mut W, |
| ) -> Result<()> { |
| let _lock = GLOBAL_OUTPUT_LOCK.lock().await; |
| while !bytes.is_empty() { |
| let count = final_stream.write(bytes).await?; |
| if count == 0 { |
| bail!("Failed to write to final stream as it was closed"); |
| } |
| bytes = &bytes[count..]; |
| } |
| Ok(()) |
| } |
|
|
| async fn write_source_mapped_final<W: AsyncWrite + Unpin>( |
| bytes: &[u8], |
| assets_for_source_mapping: Vc<AssetsForSourceMapping>, |
| root: FileSystemPath, |
| project_dir: FileSystemPath, |
| final_stream: &mut W, |
| ) -> Result<()> { |
| if let Ok(text) = std::str::from_utf8(bytes) { |
| let text = unmangle_identifiers(text, |content| { |
| format!("{{{content}}}").italic().to_string() |
| }); |
| match apply_source_mapping( |
| text.as_ref(), |
| assets_for_source_mapping, |
| root, |
| project_dir, |
| FormattingMode::AnsiColors, |
| ) |
| .await |
| { |
| Err(e) => { |
| write_final( |
| format!("Error applying source mapping: {e}\n").as_bytes(), |
| final_stream, |
| ) |
| .await?; |
| write_final(text.as_bytes(), final_stream).await?; |
| } |
| Ok(text) => { |
| write_final(text.as_bytes(), final_stream).await?; |
| } |
| } |
| } else { |
| write_final(bytes, final_stream).await?; |
| } |
| Ok(()) |
| } |
|
|
| let mut buffer = Vec::new(); |
| let mut own_output = FxHashMap::default(); |
| let mut nesting: u32 = 0; |
| let mut in_stack = None; |
| let mut stack_trace_buffer = Vec::new(); |
| loop { |
| let start = buffer.len(); |
| if stream |
| .read_until(b'\n', &mut buffer) |
| .await |
| .context("error reading from stream")? |
| == 0 |
| { |
| bail!("stream closed unexpectedly") |
| } |
| if buffer.len() - start == MARKER.len() + 2 |
| && &buffer[start..buffer.len() - 2] == MARKER |
| { |
| |
| buffer.pop(); |
| |
| match buffer.pop() { |
| Some(b'B') => { |
| stack_trace_buffer.clear(); |
| buffer.truncate(start); |
| nesting += 1; |
| in_stack = None; |
| continue; |
| } |
| Some(b'E') => { |
| buffer.truncate(start); |
| if let Some(in_stack) = in_stack { |
| if nesting != 0 { |
| stack_trace_buffer = buffer[in_stack..].to_vec(); |
| } |
| buffer.truncate(in_stack); |
| } |
| nesting = nesting.saturating_sub(1); |
| in_stack = None; |
| if nesting == 0 { |
| let line = Arc::from(take(&mut buffer).into_boxed_slice()); |
| let stack_trace = if stack_trace_buffer.is_empty() { |
| None |
| } else { |
| Some(Arc::from(take(&mut stack_trace_buffer).into_boxed_slice())) |
| }; |
| let entry = OutputEntry { |
| data: line, |
| stack_trace, |
| }; |
| let occurrence_number = *own_output |
| .entry(entry.clone()) |
| .and_modify(|c| *c += 1) |
| .or_insert(0); |
| let new_entry = { |
| let mut shared = shared.lock(); |
| shared.insert((entry.clone(), occurrence_number)) |
| }; |
| if !new_entry { |
| |
| |
| continue; |
| } |
| write_source_mapped_final( |
| &entry.data, |
| **assets_for_source_mapping, |
| root.clone(), |
| project_dir.clone(), |
| final_stream, |
| ) |
| .await?; |
| } |
| } |
| Some(b'S') => { |
| buffer.truncate(start); |
| in_stack = Some(start); |
| continue; |
| } |
| Some(b'D') => { |
| |
| break; |
| } |
| _ => {} |
| } |
| } |
| if nesting != 0 { |
| |
| continue; |
| } |
|
|
| write_source_mapped_final( |
| &buffer, |
| **assets_for_source_mapping, |
| root.clone(), |
| project_dir.clone(), |
| final_stream, |
| ) |
| .await?; |
| buffer.clear(); |
| } |
| Ok(()) |
| } |
| } |
|
|
| impl NodeJsPoolProcess { |
| async fn new( |
| cwd: &Path, |
| env: &FxHashMap<RcStr, RcStr>, |
| entrypoint: &Path, |
| assets_for_source_mapping: ResolvedVc<AssetsForSourceMapping>, |
| assets_root: FileSystemPath, |
| project_dir: FileSystemPath, |
| shared_stdout: SharedOutputSet, |
| shared_stderr: SharedOutputSet, |
| debug: bool, |
| ) -> Result<Self> { |
| let guard = Box::new(duration_span!("Node.js process startup")); |
| let listener = TcpListener::bind("127.0.0.1:0") |
| .await |
| .context("binding to a port")?; |
| let port = listener.local_addr().context("getting port")?.port(); |
| let mut cmd = Command::new("node"); |
| cmd.current_dir(cwd); |
| if debug { |
| cmd.arg("--inspect-brk"); |
| } |
| cmd.arg(entrypoint); |
| cmd.arg(port.to_string()); |
| cmd.env_clear(); |
| cmd.env( |
| "PATH", |
| std::env::var("PATH").expect("the PATH environment variable should always be set"), |
| ); |
| #[cfg(target_family = "windows")] |
| cmd.env( |
| "SystemRoot", |
| std::env::var("SystemRoot") |
| .expect("the SystemRoot environment variable should always be set"), |
| ); |
| cmd.envs(env); |
| cmd.stderr(Stdio::piped()); |
| cmd.stdout(Stdio::piped()); |
| cmd.kill_on_drop(true); |
|
|
| let mut child = cmd.spawn().context("spawning node pooled process")?; |
|
|
| let timeout = if debug { |
| Duration::MAX |
| } else { |
| CONNECT_TIMEOUT |
| }; |
|
|
| async fn get_output(child: &mut Child) -> Result<(String, String)> { |
| let mut stdout = Vec::new(); |
| let mut stderr = Vec::new(); |
| child |
| .stdout |
| .take() |
| .unwrap() |
| .read_to_end(&mut stdout) |
| .await?; |
| child |
| .stderr |
| .take() |
| .unwrap() |
| .read_to_end(&mut stderr) |
| .await?; |
| fn clean(buffer: Vec<u8>) -> Result<String> { |
| Ok(String::from_utf8(buffer)? |
| .lines() |
| .filter(|line| { |
| line.len() != MARKER_STR.len() + 1 && !line.starts_with(MARKER_STR) |
| }) |
| .collect::<Vec<_>>() |
| .join("\n")) |
| } |
| Ok((clean(stdout)?, clean(stderr)?)) |
| } |
|
|
| let (connection, _) = select! { |
| connection = listener.accept() => connection.context("accepting connection")?, |
| status = child.wait() => { |
| match status { |
| Ok(status) => { |
| let (stdout, stderr) = get_output(&mut child).await?; |
| bail!("node process exited before we could connect to it with {status}\nProcess output:\n{stdout}\nProcess error output:\n{stderr}"); |
| } |
| Err(err) => { |
| let _ = child.start_kill(); |
| let (stdout, stderr) = get_output(&mut child).await?; |
| bail!("node process exited before we could connect to it: {err:?}\nProcess output:\n{stdout}\nProcess error output:\n{stderr}"); |
| }, |
| } |
| }, |
| _ = sleep(timeout) => { |
| let _ = child.start_kill(); |
| let (stdout, stderr) = get_output(&mut child).await?; |
| bail!("timed out waiting for the Node.js process to connect ({timeout:?} timeout)\nProcess output:\n{stdout}\nProcess error output:\n{stderr}"); |
| }, |
| }; |
|
|
| let child_stdout = BufReader::new(child.stdout.take().unwrap()); |
| let child_stderr = BufReader::new(child.stderr.take().unwrap()); |
|
|
| let stdout_handler = OutputStreamHandler { |
| stream: child_stdout, |
| shared: shared_stdout, |
| assets_for_source_mapping, |
| root: assets_root.clone(), |
| project_dir: project_dir.clone(), |
| final_stream: stdout(), |
| }; |
| let stderr_handler = OutputStreamHandler { |
| stream: child_stderr, |
| shared: shared_stderr, |
| assets_for_source_mapping, |
| root: assets_root.clone(), |
| project_dir: project_dir.clone(), |
| final_stream: stderr(), |
| }; |
|
|
| let mut process = Self { |
| child: Some(child), |
| connection, |
| assets_for_source_mapping, |
| assets_root: assets_root.clone(), |
| project_dir: project_dir.clone(), |
| stdout_handler, |
| stderr_handler, |
| debug, |
| cpu_time_invested: Duration::ZERO, |
| }; |
|
|
| drop(guard); |
|
|
| let guard = duration_span!("Node.js initialization"); |
| let ready_signal = process.recv().await?; |
|
|
| if !ready_signal.is_empty() { |
| bail!( |
| "Node.js process didn't send the expected ready signal\nOutput:\n{}", |
| String::from_utf8_lossy(&ready_signal) |
| ); |
| } |
|
|
| drop(guard); |
|
|
| Ok(process) |
| } |
|
|
| async fn recv(&mut self) -> Result<Vec<u8>> { |
| let connection = &mut self.connection; |
| async fn with_timeout<T, E: Into<anyhow::Error>>( |
| debug: bool, |
| fast: bool, |
| future: impl Future<Output = Result<T, E>> + Send, |
| ) -> Result<T> { |
| if debug { |
| future.await.map_err(Into::into) |
| } else { |
| let time = if fast { |
| Duration::from_secs(20) |
| } else { |
| Duration::from_secs(5 * 60) |
| }; |
| timeout(time, future) |
| .await |
| .context("timeout while receiving message from process")? |
| .map_err(Into::into) |
| } |
| } |
| let debug = self.debug; |
| let recv_future = async move { |
| let packet_len = with_timeout(debug, false, connection.read_u32()) |
| .await |
| .context("reading packet length")? |
| .try_into() |
| .context("storing packet length")?; |
| let mut packet_data = vec![0; packet_len]; |
| with_timeout(debug, true, connection.read_exact(&mut packet_data)) |
| .await |
| .context("reading packet data")?; |
| Ok::<_, anyhow::Error>(packet_data) |
| }; |
| let (result, stdout, stderr) = join!( |
| recv_future, |
| self.stdout_handler.handle_operation(), |
| self.stderr_handler.handle_operation(), |
| ); |
| let result = result?; |
| stdout.context("unable to handle stdout from the Node.js process in a structured way")?; |
| stderr.context("unable to handle stderr from the Node.js process in a structured way")?; |
| Ok(result) |
| } |
|
|
| async fn send(&mut self, packet_data: Vec<u8>) -> Result<()> { |
| self.connection |
| .write_u32( |
| packet_data |
| .len() |
| .try_into() |
| .context("packet length does not fit into u32")?, |
| ) |
| .await |
| .context("writing packet length")?; |
| self.connection |
| .write_all(&packet_data) |
| .await |
| .context("writing packet data")?; |
| Ok(()) |
| } |
| } |
|
|
| #[derive(Default)] |
| struct NodeJsPoolStats { |
| pub total_bootup_time: Duration, |
| pub bootup_count: u32, |
| pub total_cold_process_time: Duration, |
| pub cold_process_count: u32, |
| pub total_warm_process_time: Duration, |
| pub warm_process_count: u32, |
| pub workers: u32, |
| pub booting_workers: u32, |
| pub queued_tasks: u32, |
| } |
|
|
| impl NodeJsPoolStats { |
| fn add_bootup_time(&mut self, time: Duration) { |
| self.total_bootup_time += time; |
| self.bootup_count += 1; |
| } |
|
|
| fn add_booting_worker(&mut self) { |
| self.booting_workers += 1; |
| self.workers += 1; |
| } |
|
|
| fn finished_booting_worker(&mut self) { |
| self.booting_workers -= 1; |
| } |
|
|
| fn remove_worker(&mut self) { |
| self.workers -= 1; |
| } |
|
|
| fn add_queued_task(&mut self) { |
| self.queued_tasks += 1; |
| } |
|
|
| fn add_cold_process_time(&mut self, time: Duration) { |
| self.total_cold_process_time += time; |
| self.cold_process_count += 1; |
| self.queued_tasks -= 1; |
| } |
|
|
| fn add_warm_process_time(&mut self, time: Duration) { |
| self.total_warm_process_time += time; |
| self.warm_process_count += 1; |
| self.queued_tasks -= 1; |
| } |
|
|
| fn estimated_bootup_time(&self) -> Duration { |
| if self.bootup_count == 0 { |
| Duration::from_millis(200) |
| } else { |
| self.total_bootup_time / self.bootup_count |
| } |
| } |
|
|
| fn estimated_warm_process_time(&self) -> Duration { |
| if self.warm_process_count == 0 { |
| self.estimated_cold_process_time() |
| } else { |
| self.total_warm_process_time / self.warm_process_count |
| } |
| } |
|
|
| fn estimated_cold_process_time(&self) -> Duration { |
| if self.cold_process_count == 0 { |
| |
| self.estimated_bootup_time() / 2 |
| } else { |
| self.total_cold_process_time / self.cold_process_count |
| } |
| } |
| fn wait_time_before_bootup(&self) -> Duration { |
| if self.workers == 0 { |
| return Duration::ZERO; |
| } |
| let booting_workers = self.booting_workers; |
| let workers = self.workers; |
| let warm_process_time = self.estimated_warm_process_time(); |
| let expected_completion = self.expected_completion(workers, booting_workers); |
|
|
| let new_process_duration = |
| self.estimated_bootup_time() + self.estimated_cold_process_time(); |
| if expected_completion + warm_process_time < new_process_duration { |
| |
| return (expected_completion + warm_process_time + new_process_duration) / 2; |
| } |
|
|
| let expected_completion_with_additional_worker = max( |
| new_process_duration, |
| self.expected_completion(workers + 1, booting_workers + 1), |
| ); |
| if expected_completion > expected_completion_with_additional_worker { |
| |
| return Duration::ZERO; |
| } |
|
|
| |
| (expected_completion + expected_completion_with_additional_worker) / 2 |
| } |
|
|
| fn expected_completion(&self, workers: u32, booting_workers: u32) -> Duration { |
| if workers == 0 { |
| return Duration::MAX; |
| } |
| let bootup_time = self.estimated_bootup_time(); |
| let cold_process_time = self.estimated_cold_process_time(); |
| let warm_process_time = self.estimated_warm_process_time(); |
| let expected_full_workers_in = booting_workers * (bootup_time / 2 + cold_process_time); |
| let expected_completed_task_until_full_workers = { |
| let millis = max(1, warm_process_time.as_millis()); |
| let ready_workers = workers - booting_workers; |
| (expected_full_workers_in.as_millis() / millis) as u32 * ready_workers |
| }; |
| let remaining_tasks = self |
| .queued_tasks |
| .saturating_sub(expected_completed_task_until_full_workers); |
| if remaining_tasks > 0 { |
| expected_full_workers_in + warm_process_time * remaining_tasks / workers |
| } else { |
| warm_process_time * self.queued_tasks / workers |
| } |
| } |
| } |
|
|
| impl Debug for NodeJsPoolStats { |
| fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { |
| f.debug_struct("NodeJsPoolStats") |
| .field("queued_tasks", &self.queued_tasks) |
| .field("workers", &self.workers) |
| .field("booting_workers", &self.booting_workers) |
| .field( |
| "expected_completion", |
| &self.expected_completion(self.workers, self.booting_workers), |
| ) |
| .field("bootup_time", &self.estimated_bootup_time()) |
| .field("cold_process_time", &self.estimated_cold_process_time()) |
| .field("warm_process_time", &self.estimated_warm_process_time()) |
| .field("bootup_count", &self.bootup_count) |
| .field("cold_process_count", &self.cold_process_count) |
| .field("warm_process_count", &self.warm_process_count) |
| .finish() |
| } |
| } |
|
|
| enum AcquiredPermits { |
| Idle { |
| |
| #[allow(dead_code)] |
| concurrency_permit: OwnedSemaphorePermit, |
| }, |
| Fresh { |
| |
| #[allow(dead_code)] |
| concurrency_permit: OwnedSemaphorePermit, |
| |
| #[allow(dead_code)] |
| bootup_permit: OwnedSemaphorePermit, |
| }, |
| } |
|
|
| type IdleProcessesList = Arc<Mutex<BinaryHeap<NodeJsPoolProcess>>>; |
|
|
| |
| |
| static ACTIVE_POOLS: Lazy<Mutex<Vec<IdleProcessesList>>> = Lazy::new(Default::default); |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| #[turbo_tasks::value(into = "new", cell = "new", serialization = "none", eq = "manual")] |
| pub struct NodeJsPool { |
| cwd: PathBuf, |
| entrypoint: PathBuf, |
| env: FxHashMap<RcStr, RcStr>, |
| pub assets_for_source_mapping: ResolvedVc<AssetsForSourceMapping>, |
| pub assets_root: FileSystemPath, |
| pub project_dir: FileSystemPath, |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| processes: Arc<Mutex<BinaryHeap<NodeJsPoolProcess>>>, |
| |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| concurrency_semaphore: Arc<Semaphore>, |
| |
| |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| bootup_semaphore: Arc<Semaphore>, |
| |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| idle_process_semaphore: Arc<Semaphore>, |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| shared_stdout: SharedOutputSet, |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| shared_stderr: SharedOutputSet, |
| debug: bool, |
| #[turbo_tasks(trace_ignore, debug_ignore)] |
| stats: Arc<Mutex<NodeJsPoolStats>>, |
| } |
|
|
| impl NodeJsPool { |
| |
| |
| pub(super) fn new( |
| cwd: PathBuf, |
| entrypoint: PathBuf, |
| env: FxHashMap<RcStr, RcStr>, |
| assets_for_source_mapping: ResolvedVc<AssetsForSourceMapping>, |
| assets_root: FileSystemPath, |
| project_dir: FileSystemPath, |
| concurrency: usize, |
| debug: bool, |
| ) -> Self { |
| Self { |
| cwd, |
| entrypoint, |
| env, |
| assets_for_source_mapping, |
| assets_root, |
| project_dir, |
| processes: Arc::new(Mutex::new(BinaryHeap::new())), |
| concurrency_semaphore: Arc::new(Semaphore::new(if debug { 1 } else { concurrency })), |
| bootup_semaphore: Arc::new(Semaphore::new(1)), |
| idle_process_semaphore: Arc::new(Semaphore::new(0)), |
| shared_stdout: Arc::new(Mutex::new(FxIndexSet::default())), |
| shared_stderr: Arc::new(Mutex::new(FxIndexSet::default())), |
| debug, |
| stats: Default::default(), |
| } |
| } |
|
|
| async fn acquire_process(&self) -> Result<(NodeJsPoolProcess, AcquiredPermits)> { |
| { |
| self.stats.lock().add_queued_task(); |
| } |
|
|
| let concurrency_permit = self.concurrency_semaphore.clone().acquire_owned().await?; |
|
|
| let bootup = async { |
| let permit = self.bootup_semaphore.clone().acquire_owned().await; |
| let wait_time = self.stats.lock().wait_time_before_bootup(); |
| tokio::time::sleep(wait_time).await; |
| permit |
| }; |
|
|
| select! { |
| idle_process_permit = self.idle_process_semaphore.clone().acquire_owned() => { |
| let idle_process_permit = idle_process_permit.context("acquiring idle process permit")?; |
| let process = { |
| let mut processes = self.processes.lock(); |
| let process = processes.pop().unwrap(); |
| if processes.is_empty() { |
| let mut pools = ACTIVE_POOLS.lock(); |
| if let Some(idx) = pools.iter().position(|p| Arc::ptr_eq(p, &self.processes)) { |
| pools.swap_remove(idx); |
| } |
| } |
| process |
| }; |
| idle_process_permit.forget(); |
| Ok((process, AcquiredPermits::Idle { concurrency_permit })) |
| }, |
| bootup_permit = bootup => { |
| let bootup_permit = bootup_permit.context("acquiring bootup permit")?; |
| { |
| self.stats.lock().add_booting_worker(); |
| } |
| let (process, bootup_time) = self.create_process().await?; |
| |
| { |
| let mut stats = self.stats.lock(); |
| stats.add_bootup_time(bootup_time); |
| stats.finished_booting_worker(); |
| } |
| |
| self.bootup_semaphore.add_permits(1); |
| Ok((process, AcquiredPermits::Fresh { concurrency_permit, bootup_permit })) |
| } |
| } |
| } |
|
|
| async fn create_process(&self) -> Result<(NodeJsPoolProcess, Duration), anyhow::Error> { |
| let start = Instant::now(); |
| let process = NodeJsPoolProcess::new( |
| self.cwd.as_path(), |
| &self.env, |
| self.entrypoint.as_path(), |
| self.assets_for_source_mapping, |
| self.assets_root.clone(), |
| self.project_dir.clone(), |
| self.shared_stdout.clone(), |
| self.shared_stderr.clone(), |
| self.debug, |
| ) |
| .await |
| .context("creating new process")?; |
| Ok((process, start.elapsed())) |
| } |
|
|
| pub async fn operation(&self) -> Result<NodeJsOperation> { |
| |
| let (process, permits) = self.acquire_process().await?; |
|
|
| Ok(NodeJsOperation { |
| process: Some(process), |
| permits, |
| processes: self.processes.clone(), |
| idle_process_semaphore: self.idle_process_semaphore.clone(), |
| start: Instant::now(), |
| stats: self.stats.clone(), |
| allow_process_reuse: true, |
| }) |
| } |
|
|
| pub fn scale_down() { |
| let pools = ACTIVE_POOLS.lock().clone(); |
| for pool in pools { |
| let mut pool = pool.lock(); |
| let best = pool.pop().unwrap(); |
| pool.clear(); |
| pool.push(best); |
| pool.shrink_to_fit(); |
| } |
| } |
|
|
| pub fn scale_zero() { |
| let pools = take(&mut *ACTIVE_POOLS.lock()); |
| for pool in pools { |
| let mut pool = pool.lock(); |
| pool.clear(); |
| pool.shrink_to_fit(); |
| } |
| } |
| } |
|
|
| pub struct NodeJsOperation { |
| process: Option<NodeJsPoolProcess>, |
| |
| #[allow(dead_code)] |
| permits: AcquiredPermits, |
| processes: Arc<Mutex<BinaryHeap<NodeJsPoolProcess>>>, |
| idle_process_semaphore: Arc<Semaphore>, |
| start: Instant, |
| stats: Arc<Mutex<NodeJsPoolStats>>, |
| allow_process_reuse: bool, |
| } |
|
|
| impl NodeJsOperation { |
| async fn with_process<'a, F: Future<Output = Result<T>> + Send + 'a, T>( |
| &'a mut self, |
| f: impl FnOnce(&'a mut NodeJsPoolProcess) -> F, |
| ) -> Result<T> { |
| let process = self |
| .process |
| .as_mut() |
| .context("Node.js operation already finished")?; |
|
|
| if !self.allow_process_reuse { |
| bail!("Node.js process is no longer usable"); |
| } |
|
|
| let result = f(process).await; |
| if result.is_err() && self.allow_process_reuse { |
| self.stats.lock().remove_worker(); |
| self.allow_process_reuse = false; |
| } |
| result |
| } |
|
|
| pub async fn recv<M>(&mut self) -> Result<M> |
| where |
| M: DeserializeOwned, |
| { |
| let message = self |
| .with_process(|process| async move { |
| process.recv().await.context("failed to receive message") |
| }) |
| .await?; |
| let message = std::str::from_utf8(&message).context("message is not valid UTF-8")?; |
| parse_json_with_source_context(message).context("failed to deserialize message") |
| } |
|
|
| pub async fn send<M>(&mut self, message: M) -> Result<()> |
| where |
| M: Serialize, |
| { |
| let message = serde_json::to_vec(&message).context("failed to serialize message")?; |
| self.with_process(|process| async move { |
| timeout(Duration::from_secs(30), process.send(message)) |
| .await |
| .context("timeout while sending message")? |
| .context("failed to send message")?; |
| Ok(()) |
| }) |
| .await |
| } |
|
|
| pub async fn wait_or_kill(mut self) -> Result<ExitStatus> { |
| let mut process = self |
| .process |
| .take() |
| .context("Node.js operation already finished")?; |
|
|
| if self.allow_process_reuse { |
| self.stats.lock().remove_worker(); |
| } |
|
|
| let mut child = process |
| .child |
| .take() |
| .context("Node.js operation already finished")?; |
|
|
| |
| let _ = child.start_kill(); |
| let status = timeout(Duration::from_secs(30), child.wait()) |
| .await |
| .context("timeout while waiting for process end")? |
| .context("waiting for process end")?; |
|
|
| Ok(status) |
| } |
|
|
| pub fn disallow_reuse(&mut self) { |
| if self.allow_process_reuse { |
| self.stats.lock().remove_worker(); |
| self.allow_process_reuse = false; |
| } |
| } |
|
|
| pub async fn apply_source_mapping<'a>( |
| &self, |
| text: &'a str, |
| formatting_mode: FormattingMode, |
| ) -> Result<Cow<'a, str>> { |
| if let Some(process) = self.process.as_ref() { |
| process.apply_source_mapping(text, formatting_mode).await |
| } else { |
| Ok(Cow::Borrowed(text)) |
| } |
| } |
| } |
|
|
| impl Drop for NodeJsOperation { |
| fn drop(&mut self) { |
| if let Some(mut process) = self.process.take() { |
| let elapsed = self.start.elapsed(); |
| { |
| let stats = &mut self.stats.lock(); |
| match self.permits { |
| AcquiredPermits::Idle { .. } => stats.add_warm_process_time(elapsed), |
| AcquiredPermits::Fresh { .. } => stats.add_cold_process_time(elapsed), |
| } |
| } |
| if self.allow_process_reuse { |
| process.cpu_time_invested += elapsed; |
| { |
| let mut processes = self.processes.lock(); |
| if processes.is_empty() { |
| ACTIVE_POOLS.lock().push(self.processes.clone()); |
| } |
| processes.push(process); |
| } |
| self.idle_process_semaphore.add_permits(1); |
| } |
| } |
| } |
| } |
|
|