mirror of
https://github.com/rust-lang/rust-analyzer
synced 2024-12-29 06:23:25 +00:00
Auto merge of #14967 - rust-lang:revert-14965-panic-ctx, r=Veykril
Revert "Add mandatory panic contexts to all threadpool tasks" Reverts rust-lang/rust-analyzer#14965 This won't quite work actually given the use of `catch_unwind` in some of these
This commit is contained in:
commit
48f8799bf1
5 changed files with 138 additions and 174 deletions
|
@ -104,10 +104,13 @@ impl<'a> RequestDispatcher<'a> {
|
||||||
None => return self,
|
None => return self,
|
||||||
};
|
};
|
||||||
|
|
||||||
self.global_state.task_pool.handle.spawn(ThreadIntent::Worker, panic_context, {
|
self.global_state.task_pool.handle.spawn(ThreadIntent::Worker, {
|
||||||
let world = self.global_state.snapshot();
|
let world = self.global_state.snapshot();
|
||||||
move || {
|
move || {
|
||||||
let result = panic::catch_unwind(move || f(world, params));
|
let result = panic::catch_unwind(move || {
|
||||||
|
let _pctx = stdx::panic_context::enter(panic_context);
|
||||||
|
f(world, params)
|
||||||
|
});
|
||||||
match thread_result_to_response::<R>(req.id.clone(), result) {
|
match thread_result_to_response::<R>(req.id.clone(), result) {
|
||||||
Ok(response) => Task::Response(response),
|
Ok(response) => Task::Response(response),
|
||||||
Err(_) => Task::Response(lsp_server::Response::new_err(
|
Err(_) => Task::Response(lsp_server::Response::new_err(
|
||||||
|
@ -175,10 +178,13 @@ impl<'a> RequestDispatcher<'a> {
|
||||||
None => return self,
|
None => return self,
|
||||||
};
|
};
|
||||||
|
|
||||||
self.global_state.task_pool.handle.spawn(intent, panic_context, {
|
self.global_state.task_pool.handle.spawn(intent, {
|
||||||
let world = self.global_state.snapshot();
|
let world = self.global_state.snapshot();
|
||||||
move || {
|
move || {
|
||||||
let result = panic::catch_unwind(move || f(world, params));
|
let result = panic::catch_unwind(move || {
|
||||||
|
let _pctx = stdx::panic_context::enter(panic_context);
|
||||||
|
f(world, params)
|
||||||
|
});
|
||||||
match thread_result_to_response::<R>(req.id.clone(), result) {
|
match thread_result_to_response::<R>(req.id.clone(), result) {
|
||||||
Ok(response) => Task::Response(response),
|
Ok(response) => Task::Response(response),
|
||||||
Err(_) => Task::Retry(req),
|
Err(_) => Task::Retry(req),
|
||||||
|
|
|
@ -291,15 +291,11 @@ fn run_flycheck(state: &mut GlobalState, vfs_path: VfsPath) -> bool {
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
};
|
};
|
||||||
state.task_pool.handle.spawn_with_sender(
|
state.task_pool.handle.spawn_with_sender(stdx::thread::ThreadIntent::Worker, move |_| {
|
||||||
stdx::thread::ThreadIntent::Worker,
|
if let Err(e) = std::panic::catch_unwind(task) {
|
||||||
"flycheck",
|
tracing::error!("flycheck task panicked: {e:?}")
|
||||||
move |_| {
|
}
|
||||||
if let Err(e) = std::panic::catch_unwind(task) {
|
});
|
||||||
tracing::error!("flycheck task panicked: {e:?}")
|
|
||||||
}
|
|
||||||
},
|
|
||||||
);
|
|
||||||
true
|
true
|
||||||
} else {
|
} else {
|
||||||
false
|
false
|
||||||
|
|
|
@ -397,25 +397,19 @@ impl GlobalState {
|
||||||
tracing::debug!(%cause, "will prime caches");
|
tracing::debug!(%cause, "will prime caches");
|
||||||
let num_worker_threads = self.config.prime_caches_num_threads();
|
let num_worker_threads = self.config.prime_caches_num_threads();
|
||||||
|
|
||||||
self.task_pool.handle.spawn_with_sender(
|
self.task_pool.handle.spawn_with_sender(stdx::thread::ThreadIntent::Worker, {
|
||||||
stdx::thread::ThreadIntent::Worker,
|
let analysis = self.snapshot().analysis;
|
||||||
"prime_caches",
|
move |sender| {
|
||||||
{
|
sender.send(Task::PrimeCaches(PrimeCachesProgress::Begin)).unwrap();
|
||||||
let analysis = self.snapshot().analysis;
|
let res = analysis.parallel_prime_caches(num_worker_threads, |progress| {
|
||||||
move |sender| {
|
let report = PrimeCachesProgress::Report(progress);
|
||||||
sender.send(Task::PrimeCaches(PrimeCachesProgress::Begin)).unwrap();
|
sender.send(Task::PrimeCaches(report)).unwrap();
|
||||||
let res = analysis.parallel_prime_caches(num_worker_threads, |progress| {
|
});
|
||||||
let report = PrimeCachesProgress::Report(progress);
|
sender
|
||||||
sender.send(Task::PrimeCaches(report)).unwrap();
|
.send(Task::PrimeCaches(PrimeCachesProgress::End { cancelled: res.is_err() }))
|
||||||
});
|
.unwrap();
|
||||||
sender
|
}
|
||||||
.send(Task::PrimeCaches(PrimeCachesProgress::End {
|
});
|
||||||
cancelled: res.is_err(),
|
|
||||||
}))
|
|
||||||
.unwrap();
|
|
||||||
}
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn update_status_or_notify(&mut self) {
|
fn update_status_or_notify(&mut self) {
|
||||||
|
@ -802,62 +796,56 @@ impl GlobalState {
|
||||||
|
|
||||||
// Diagnostics are triggered by the user typing
|
// Diagnostics are triggered by the user typing
|
||||||
// so we run them on a latency sensitive thread.
|
// so we run them on a latency sensitive thread.
|
||||||
self.task_pool.handle.spawn(
|
self.task_pool.handle.spawn(stdx::thread::ThreadIntent::LatencySensitive, move || {
|
||||||
stdx::thread::ThreadIntent::LatencySensitive,
|
let _p = profile::span("publish_diagnostics");
|
||||||
"publish_diagnostics",
|
let diagnostics = subscriptions
|
||||||
move || {
|
.into_iter()
|
||||||
let _p = profile::span("publish_diagnostics");
|
.filter_map(|file_id| {
|
||||||
let diagnostics = subscriptions
|
let line_index = snapshot.file_line_index(file_id).ok()?;
|
||||||
.into_iter()
|
Some((
|
||||||
.filter_map(|file_id| {
|
file_id,
|
||||||
let line_index = snapshot.file_line_index(file_id).ok()?;
|
line_index,
|
||||||
Some((
|
snapshot
|
||||||
file_id,
|
.analysis
|
||||||
line_index,
|
.diagnostics(
|
||||||
snapshot
|
&snapshot.config.diagnostics(),
|
||||||
.analysis
|
ide::AssistResolveStrategy::None,
|
||||||
.diagnostics(
|
file_id,
|
||||||
&snapshot.config.diagnostics(),
|
)
|
||||||
ide::AssistResolveStrategy::None,
|
.ok()?,
|
||||||
file_id,
|
))
|
||||||
)
|
})
|
||||||
.ok()?,
|
.map(|(file_id, line_index, it)| {
|
||||||
))
|
(
|
||||||
})
|
file_id,
|
||||||
.map(|(file_id, line_index, it)| {
|
it.into_iter()
|
||||||
(
|
.map(move |d| lsp_types::Diagnostic {
|
||||||
file_id,
|
range: crate::to_proto::range(&line_index, d.range),
|
||||||
it.into_iter()
|
severity: Some(crate::to_proto::diagnostic_severity(d.severity)),
|
||||||
.map(move |d| lsp_types::Diagnostic {
|
code: Some(lsp_types::NumberOrString::String(
|
||||||
range: crate::to_proto::range(&line_index, d.range),
|
d.code.as_str().to_string(),
|
||||||
severity: Some(crate::to_proto::diagnostic_severity(
|
)),
|
||||||
d.severity,
|
code_description: Some(lsp_types::CodeDescription {
|
||||||
)),
|
href: lsp_types::Url::parse(&format!(
|
||||||
code: Some(lsp_types::NumberOrString::String(
|
"https://rust-analyzer.github.io/manual.html#{}",
|
||||||
d.code.as_str().to_string(),
|
d.code.as_str()
|
||||||
)),
|
))
|
||||||
code_description: Some(lsp_types::CodeDescription {
|
.unwrap(),
|
||||||
href: lsp_types::Url::parse(&format!(
|
}),
|
||||||
"https://rust-analyzer.github.io/manual.html#{}",
|
source: Some("rust-analyzer".to_string()),
|
||||||
d.code.as_str()
|
message: d.message,
|
||||||
))
|
related_information: None,
|
||||||
.unwrap(),
|
tags: if d.unused {
|
||||||
}),
|
Some(vec![lsp_types::DiagnosticTag::UNNECESSARY])
|
||||||
source: Some("rust-analyzer".to_string()),
|
} else {
|
||||||
message: d.message,
|
None
|
||||||
related_information: None,
|
},
|
||||||
tags: if d.unused {
|
data: None,
|
||||||
Some(vec![lsp_types::DiagnosticTag::UNNECESSARY])
|
})
|
||||||
} else {
|
.collect::<Vec<_>>(),
|
||||||
None
|
)
|
||||||
},
|
});
|
||||||
data: None,
|
Task::Diagnostics(diagnostics.collect())
|
||||||
})
|
});
|
||||||
.collect::<Vec<_>>(),
|
|
||||||
)
|
|
||||||
});
|
|
||||||
Task::Diagnostics(diagnostics.collect())
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -185,7 +185,7 @@ impl GlobalState {
|
||||||
pub(crate) fn fetch_workspaces(&mut self, cause: Cause) {
|
pub(crate) fn fetch_workspaces(&mut self, cause: Cause) {
|
||||||
tracing::info!(%cause, "will fetch workspaces");
|
tracing::info!(%cause, "will fetch workspaces");
|
||||||
|
|
||||||
self.task_pool.handle.spawn_with_sender(ThreadIntent::Worker, "fetch_workspaces", {
|
self.task_pool.handle.spawn_with_sender(ThreadIntent::Worker, {
|
||||||
let linked_projects = self.config.linked_projects();
|
let linked_projects = self.config.linked_projects();
|
||||||
let detached_files = self.config.detached_files().to_vec();
|
let detached_files = self.config.detached_files().to_vec();
|
||||||
let cargo_config = self.config.cargo();
|
let cargo_config = self.config.cargo();
|
||||||
|
@ -260,25 +260,19 @@ impl GlobalState {
|
||||||
tracing::info!(%cause, "will fetch build data");
|
tracing::info!(%cause, "will fetch build data");
|
||||||
let workspaces = Arc::clone(&self.workspaces);
|
let workspaces = Arc::clone(&self.workspaces);
|
||||||
let config = self.config.cargo();
|
let config = self.config.cargo();
|
||||||
self.task_pool.handle.spawn_with_sender(
|
self.task_pool.handle.spawn_with_sender(ThreadIntent::Worker, move |sender| {
|
||||||
ThreadIntent::Worker,
|
sender.send(Task::FetchBuildData(BuildDataProgress::Begin)).unwrap();
|
||||||
"fetch_build_data",
|
|
||||||
move |sender| {
|
|
||||||
sender.send(Task::FetchBuildData(BuildDataProgress::Begin)).unwrap();
|
|
||||||
|
|
||||||
let progress = {
|
let progress = {
|
||||||
let sender = sender.clone();
|
let sender = sender.clone();
|
||||||
move |msg| {
|
move |msg| {
|
||||||
sender.send(Task::FetchBuildData(BuildDataProgress::Report(msg))).unwrap()
|
sender.send(Task::FetchBuildData(BuildDataProgress::Report(msg))).unwrap()
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
let res = ProjectWorkspace::run_all_build_scripts(&workspaces, &config, &progress);
|
let res = ProjectWorkspace::run_all_build_scripts(&workspaces, &config, &progress);
|
||||||
|
|
||||||
sender
|
sender.send(Task::FetchBuildData(BuildDataProgress::End((workspaces, res)))).unwrap();
|
||||||
.send(Task::FetchBuildData(BuildDataProgress::End((workspaces, res))))
|
});
|
||||||
.unwrap();
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn fetch_proc_macros(&mut self, cause: Cause, paths: Vec<ProcMacroPaths>) {
|
pub(crate) fn fetch_proc_macros(&mut self, cause: Cause, paths: Vec<ProcMacroPaths>) {
|
||||||
|
@ -286,54 +280,50 @@ impl GlobalState {
|
||||||
let dummy_replacements = self.config.dummy_replacements().clone();
|
let dummy_replacements = self.config.dummy_replacements().clone();
|
||||||
let proc_macro_clients = self.proc_macro_clients.clone();
|
let proc_macro_clients = self.proc_macro_clients.clone();
|
||||||
|
|
||||||
self.task_pool.handle.spawn_with_sender(
|
self.task_pool.handle.spawn_with_sender(ThreadIntent::Worker, move |sender| {
|
||||||
ThreadIntent::Worker,
|
sender.send(Task::LoadProcMacros(ProcMacroProgress::Begin)).unwrap();
|
||||||
"fetch_proc_macros",
|
|
||||||
move |sender| {
|
|
||||||
sender.send(Task::LoadProcMacros(ProcMacroProgress::Begin)).unwrap();
|
|
||||||
|
|
||||||
let dummy_replacements = &dummy_replacements;
|
let dummy_replacements = &dummy_replacements;
|
||||||
let progress = {
|
let progress = {
|
||||||
let sender = sender.clone();
|
let sender = sender.clone();
|
||||||
&move |msg| {
|
&move |msg| {
|
||||||
sender.send(Task::LoadProcMacros(ProcMacroProgress::Report(msg))).unwrap()
|
sender.send(Task::LoadProcMacros(ProcMacroProgress::Report(msg))).unwrap()
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut res = FxHashMap::default();
|
|
||||||
let chain = proc_macro_clients
|
|
||||||
.iter()
|
|
||||||
.map(|res| res.as_ref().map_err(|e| e.to_string()))
|
|
||||||
.chain(iter::repeat_with(|| Err("Proc macros servers are not running".into())));
|
|
||||||
for (client, paths) in chain.zip(paths) {
|
|
||||||
res.extend(paths.into_iter().map(move |(crate_id, res)| {
|
|
||||||
(
|
|
||||||
crate_id,
|
|
||||||
res.map_or_else(
|
|
||||||
|_| Err("proc macro crate is missing dylib".to_owned()),
|
|
||||||
|(crate_name, path)| {
|
|
||||||
progress(path.display().to_string());
|
|
||||||
client.as_ref().map_err(Clone::clone).and_then(|client| {
|
|
||||||
load_proc_macro(
|
|
||||||
client,
|
|
||||||
&path,
|
|
||||||
crate_name
|
|
||||||
.as_deref()
|
|
||||||
.and_then(|crate_name| {
|
|
||||||
dummy_replacements.get(crate_name).map(|v| &**v)
|
|
||||||
})
|
|
||||||
.unwrap_or_default(),
|
|
||||||
)
|
|
||||||
})
|
|
||||||
},
|
|
||||||
),
|
|
||||||
)
|
|
||||||
}));
|
|
||||||
}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
sender.send(Task::LoadProcMacros(ProcMacroProgress::End(res))).unwrap();
|
let mut res = FxHashMap::default();
|
||||||
},
|
let chain = proc_macro_clients
|
||||||
);
|
.iter()
|
||||||
|
.map(|res| res.as_ref().map_err(|e| e.to_string()))
|
||||||
|
.chain(iter::repeat_with(|| Err("Proc macros servers are not running".into())));
|
||||||
|
for (client, paths) in chain.zip(paths) {
|
||||||
|
res.extend(paths.into_iter().map(move |(crate_id, res)| {
|
||||||
|
(
|
||||||
|
crate_id,
|
||||||
|
res.map_or_else(
|
||||||
|
|_| Err("proc macro crate is missing dylib".to_owned()),
|
||||||
|
|(crate_name, path)| {
|
||||||
|
progress(path.display().to_string());
|
||||||
|
client.as_ref().map_err(Clone::clone).and_then(|client| {
|
||||||
|
load_proc_macro(
|
||||||
|
client,
|
||||||
|
&path,
|
||||||
|
crate_name
|
||||||
|
.as_deref()
|
||||||
|
.and_then(|crate_name| {
|
||||||
|
dummy_replacements.get(crate_name).map(|v| &**v)
|
||||||
|
})
|
||||||
|
.unwrap_or_default(),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
},
|
||||||
|
),
|
||||||
|
)
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
|
sender.send(Task::LoadProcMacros(ProcMacroProgress::End(res))).unwrap();
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn set_proc_macros(&mut self, proc_macros: ProcMacros) {
|
pub(crate) fn set_proc_macros(&mut self, proc_macros: ProcMacros) {
|
||||||
|
|
|
@ -14,41 +14,25 @@ impl<T> TaskPool<T> {
|
||||||
TaskPool { sender, pool: Pool::new(threads) }
|
TaskPool { sender, pool: Pool::new(threads) }
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn spawn<F>(
|
pub(crate) fn spawn<F>(&mut self, intent: ThreadIntent, task: F)
|
||||||
&mut self,
|
where
|
||||||
intent: ThreadIntent,
|
|
||||||
panic_context: impl Into<String>,
|
|
||||||
task: F,
|
|
||||||
) where
|
|
||||||
F: FnOnce() -> T + Send + 'static,
|
F: FnOnce() -> T + Send + 'static,
|
||||||
T: Send + 'static,
|
T: Send + 'static,
|
||||||
{
|
{
|
||||||
let panic_context = panic_context.into();
|
|
||||||
self.pool.spawn(intent, {
|
self.pool.spawn(intent, {
|
||||||
let sender = self.sender.clone();
|
let sender = self.sender.clone();
|
||||||
move || {
|
move || sender.send(task()).unwrap()
|
||||||
let _pctx = stdx::panic_context::enter(panic_context);
|
|
||||||
sender.send(task()).unwrap()
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn spawn_with_sender<F>(
|
pub(crate) fn spawn_with_sender<F>(&mut self, intent: ThreadIntent, task: F)
|
||||||
&mut self,
|
where
|
||||||
intent: ThreadIntent,
|
|
||||||
panic_context: impl Into<String>,
|
|
||||||
task: F,
|
|
||||||
) where
|
|
||||||
F: FnOnce(Sender<T>) + Send + 'static,
|
F: FnOnce(Sender<T>) + Send + 'static,
|
||||||
T: Send + 'static,
|
T: Send + 'static,
|
||||||
{
|
{
|
||||||
let panic_context = panic_context.into();
|
|
||||||
self.pool.spawn(intent, {
|
self.pool.spawn(intent, {
|
||||||
let sender = self.sender.clone();
|
let sender = self.sender.clone();
|
||||||
move || {
|
move || task(sender)
|
||||||
let _pctx = stdx::panic_context::enter(panic_context);
|
|
||||||
task(sender)
|
|
||||||
}
|
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue