mirror of
https://github.com/DioxusLabs/dioxus
synced 2024-11-26 14:10:20 +00:00
Try to rerun all dirty scopes before polling any tasks to fix effect ordering
This commit is contained in:
parent
47e46de74f
commit
701093ede5
4 changed files with 92 additions and 30 deletions
|
@ -14,22 +14,39 @@ fn app() -> Element {
|
|||
rsx! {
|
||||
div {
|
||||
button { onclick: move |_| counters.write().push(0), "Add counter" }
|
||||
button { onclick: move |_| { counters.write().pop(); }, "Remove counter" }
|
||||
button {
|
||||
onclick: move |_| {
|
||||
counters.write().pop();
|
||||
},
|
||||
"Remove counter"
|
||||
}
|
||||
p { "Total: {sum}" }
|
||||
for (i, counter) in counters.read().iter().enumerate() {
|
||||
li {
|
||||
button { onclick: move |_| counters.write()[i] -= 1, "-1" }
|
||||
input {
|
||||
value: "{counter}",
|
||||
oninput: move |e| {
|
||||
if let Ok(value) = e.value().parse::<usize>() {
|
||||
counters.write()[i] = value;
|
||||
}
|
||||
}
|
||||
}
|
||||
button { onclick: move |_| counters.write()[i] += 1, "+1" }
|
||||
button { onclick: move |_| { counters.write().remove(i); }, "x" }
|
||||
}
|
||||
for i in 0..counters.len() {
|
||||
Child { i, counters }
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[component]
|
||||
fn Child(i: usize, counters: Signal<Vec<usize>>) -> Element {
|
||||
rsx! {
|
||||
li {
|
||||
button { onclick: move |_| counters.write()[i] -= 1, "-1" }
|
||||
input {
|
||||
value: "{counters.read()[i]}",
|
||||
oninput: move |e| {
|
||||
if let Ok(value) = e.value().parse::<usize>() {
|
||||
counters.write()[i] = value;
|
||||
}
|
||||
}
|
||||
}
|
||||
button { onclick: move |_| counters.write()[i] += 1, "+1" }
|
||||
button {
|
||||
onclick: move |_| {
|
||||
counters.write().remove(i);
|
||||
},
|
||||
"x"
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -68,12 +68,18 @@ impl VirtualDom {
|
|||
//
|
||||
// Note: This will not remove any ids from the arena
|
||||
pub(crate) fn drop_scope(&mut self, id: ScopeId) {
|
||||
self.dirty_scopes.remove(&DirtyScope {
|
||||
height: self.scopes[id.0].context().height,
|
||||
id,
|
||||
});
|
||||
let (height, spawned_tasks) = {
|
||||
let scope = self.scopes.remove(id.0);
|
||||
let context = scope.context();
|
||||
let spawned_tasks = context.spawned_tasks.borrow();
|
||||
let spawned_tasks: Vec<_> = spawned_tasks.iter().copied().collect();
|
||||
(context.height, spawned_tasks)
|
||||
};
|
||||
|
||||
self.scopes.remove(id.0);
|
||||
self.dirty_scopes.remove(&DirtyScope { height, id });
|
||||
for task in spawned_tasks {
|
||||
self.runtime.remove_task(task);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,12 @@ use crate::{
|
|||
use futures_util::{pin_mut, StreamExt};
|
||||
use rustc_hash::{FxHashMap, FxHashSet};
|
||||
use slab::Slab;
|
||||
use std::{any::Any, collections::BTreeSet, future::Future, rc::Rc};
|
||||
use std::{
|
||||
any::Any,
|
||||
collections::{BTreeSet, VecDeque},
|
||||
future::Future,
|
||||
rc::Rc,
|
||||
};
|
||||
|
||||
/// A virtual node system that progresses user events and diffs UI trees.
|
||||
///
|
||||
|
@ -184,6 +189,7 @@ pub struct VirtualDom {
|
|||
pub(crate) scopes: Slab<ScopeState>,
|
||||
|
||||
pub(crate) dirty_scopes: BTreeSet<DirtyScope>,
|
||||
pub(crate) dirty_tasks: VecDeque<Task>,
|
||||
|
||||
// Maps a template path to a map of byte indexes to templates
|
||||
pub(crate) templates: FxHashMap<TemplateId, FxHashMap<usize, Template>>,
|
||||
|
@ -227,7 +233,7 @@ impl VirtualDom {
|
|||
///
|
||||
/// Note: the VirtualDom is not progressed, you must either "run_with_deadline" or use "rebuild" to progress it.
|
||||
pub fn new(app: fn() -> Element) -> Self {
|
||||
Self::new_with_props(move || app(), ())
|
||||
Self::new_with_props(app, ())
|
||||
}
|
||||
|
||||
/// Create a new virtualdom and build it immediately
|
||||
|
@ -278,6 +284,7 @@ impl VirtualDom {
|
|||
runtime: Runtime::new(tx),
|
||||
scopes: Default::default(),
|
||||
dirty_scopes: Default::default(),
|
||||
dirty_tasks: Default::default(),
|
||||
templates: Default::default(),
|
||||
queued_templates: Default::default(),
|
||||
elements: Default::default(),
|
||||
|
@ -399,9 +406,8 @@ impl VirtualDom {
|
|||
if let Some(msg) = some_msg.take() {
|
||||
match msg {
|
||||
SchedulerMsg::Immediate(id) => self.mark_dirty(id),
|
||||
SchedulerMsg::TaskNotified(task) => self.handle_task_wakeup(task),
|
||||
SchedulerMsg::TaskNotified(task) => self.queue_task_wakeup(task),
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
// If they're not ready, then we should wait for them to be ready
|
||||
|
@ -409,8 +415,18 @@ impl VirtualDom {
|
|||
Ok(Some(val)) => some_msg = Some(val),
|
||||
Ok(None) => return,
|
||||
Err(_) => {
|
||||
// Now that we have collected all queued work, we should check if we have any dirty scopes. If there are not, then we can poll any queued futures
|
||||
let has_dirty_scopes = !self.dirty_scopes.is_empty();
|
||||
|
||||
if !has_dirty_scopes {
|
||||
// If we have no dirty scopes, then we should poll any tasks that have been notified
|
||||
while let Some(task) = self.dirty_tasks.pop_front() {
|
||||
self.handle_task_wakeup(task);
|
||||
}
|
||||
}
|
||||
|
||||
// If we have any dirty scopes, or finished fiber trees then we should exit
|
||||
if !self.dirty_scopes.is_empty() || !self.suspended_scopes.is_empty() {
|
||||
if has_dirty_scopes || !self.suspended_scopes.is_empty() {
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -425,7 +441,7 @@ impl VirtualDom {
|
|||
while let Ok(Some(msg)) = self.rx.try_next() {
|
||||
match msg {
|
||||
SchedulerMsg::Immediate(id) => self.mark_dirty(id),
|
||||
SchedulerMsg::TaskNotified(task) => self.handle_task_wakeup(task),
|
||||
SchedulerMsg::TaskNotified(task) => self.queue_task_wakeup(task),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -608,6 +624,11 @@ impl VirtualDom {
|
|||
}
|
||||
}
|
||||
|
||||
/// Queue a task to be polled after all dirty scopes have been rendered
|
||||
fn queue_task_wakeup(&mut self, id: Task) {
|
||||
self.dirty_tasks.push_back(id);
|
||||
}
|
||||
|
||||
/// Handle notifications by tasks inside the scheduler
|
||||
///
|
||||
/// This is precise, meaning we won't poll every task, just tasks that have woken up as notified to use by the
|
||||
|
|
|
@ -10,7 +10,7 @@ use std::{
|
|||
};
|
||||
|
||||
macro_rules! read_impls {
|
||||
($ty:ident, $bound:path) => {
|
||||
($ty:ident, $bound:path, $vec_bound:path) => {
|
||||
impl<T: Default + 'static, S: $bound> Default for $ty<T, S> {
|
||||
#[track_caller]
|
||||
fn default() -> Self {
|
||||
|
@ -47,6 +47,20 @@ macro_rules! read_impls {
|
|||
self.with(|v| *v == *other)
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: 'static, S: $vec_bound> $ty<Vec<T>, S> {
|
||||
/// Returns the length of the inner vector.
|
||||
#[track_caller]
|
||||
pub fn len(&self) -> usize {
|
||||
self.with(|v| v.len())
|
||||
}
|
||||
|
||||
/// Returns true if the inner vector is empty.
|
||||
#[track_caller]
|
||||
pub fn is_empty(&self) -> bool {
|
||||
self.with(|v| v.is_empty())
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
@ -180,7 +194,7 @@ macro_rules! write_impls {
|
|||
};
|
||||
}
|
||||
|
||||
read_impls!(CopyValue, Storage<T>);
|
||||
read_impls!(CopyValue, Storage<T>, Storage<Vec<T>>);
|
||||
|
||||
impl<T: 'static, S: Storage<Vec<T>>> CopyValue<Vec<T>, S> {
|
||||
/// Read a value from the inner vector.
|
||||
|
@ -245,7 +259,7 @@ impl<T: 'static, S: Storage<Option<T>>> CopyValue<Option<T>, S> {
|
|||
}
|
||||
}
|
||||
|
||||
read_impls!(Signal, Storage<SignalData<T>>);
|
||||
read_impls!(Signal, Storage<SignalData<T>>, Storage<SignalData<Vec<T>>>);
|
||||
|
||||
impl<T: 'static, S: Storage<SignalData<Vec<T>>>> Signal<Vec<T>, S> {
|
||||
/// Read a value from the inner vector.
|
||||
|
@ -323,7 +337,11 @@ impl<T: 'static, S: Storage<SignalData<Option<T>>>> Signal<Option<T>, S> {
|
|||
}
|
||||
}
|
||||
|
||||
read_impls!(ReadOnlySignal, Storage<SignalData<T>>);
|
||||
read_impls!(
|
||||
ReadOnlySignal,
|
||||
Storage<SignalData<T>>,
|
||||
Storage<SignalData<Vec<T>>>
|
||||
);
|
||||
|
||||
/// An iterator over the values of a `CopyValue<Vec<T>>`.
|
||||
pub struct CopyValueIterator<T: 'static, S: Storage<Vec<T>>> {
|
||||
|
|
Loading…
Reference in a new issue