mirror of
https://github.com/servo/servo.git
synced 2025-08-03 04:30:10 +01:00
Reuse the prev_sibling slot for free_count to save a word.
MozReview-Commit-ID: 9jVkDM4P8mC
This commit is contained in:
parent
65fc5a9bd5
commit
e93b7fb3e8
2 changed files with 46 additions and 21 deletions
|
@ -498,6 +498,25 @@ impl CascadeLevel {
|
|||
}
|
||||
}
|
||||
|
||||
// The root node never has siblings, but needs a free count. We use the same
|
||||
// storage for both to save memory.
|
||||
struct PrevSiblingOrFreeCount(AtomicPtr<RuleNode>);
|
||||
impl PrevSiblingOrFreeCount {
|
||||
fn new() -> Self {
|
||||
PrevSiblingOrFreeCount(AtomicPtr::new(ptr::null_mut()))
|
||||
}
|
||||
|
||||
unsafe fn as_prev_sibling(&self) -> &AtomicPtr<RuleNode> {
|
||||
&self.0
|
||||
}
|
||||
|
||||
unsafe fn as_free_count(&self) -> &AtomicUsize {
|
||||
unsafe {
|
||||
mem::transmute(&self.0)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// A node in the rule tree.
|
||||
pub struct RuleNode {
|
||||
/// The root node. Only the root has no root pointer, for obvious reasons.
|
||||
|
@ -516,18 +535,16 @@ pub struct RuleNode {
|
|||
refcount: AtomicUsize,
|
||||
first_child: AtomicPtr<RuleNode>,
|
||||
next_sibling: AtomicPtr<RuleNode>,
|
||||
prev_sibling: AtomicPtr<RuleNode>,
|
||||
|
||||
/// Previous sibling pointer for all non-root nodes.
|
||||
///
|
||||
/// For the root, stores the of RuleNodes we have added to the free list
|
||||
/// since the last GC. (We don't update this if we rescue a RuleNode from
|
||||
/// the free list. It's just used as a heuristic to decide when to run GC.)
|
||||
prev_sibling_or_free_count: PrevSiblingOrFreeCount,
|
||||
|
||||
/// The next item in the rule tree free list, that starts on the root node.
|
||||
next_free: AtomicPtr<RuleNode>,
|
||||
|
||||
/// Number of RuleNodes we have added to the free list since the last GC.
|
||||
/// (We don't update this if we rescue a RuleNode from the free list. It's
|
||||
/// just used as a heuristic to decide when to run GC.)
|
||||
///
|
||||
/// Only used on the root RuleNode. (We could probably re-use one of the
|
||||
/// sibling pointers to save space.)
|
||||
free_count: AtomicUsize,
|
||||
}
|
||||
|
||||
unsafe impl Sync for RuleTree {}
|
||||
|
@ -547,9 +564,8 @@ impl RuleNode {
|
|||
refcount: AtomicUsize::new(1),
|
||||
first_child: AtomicPtr::new(ptr::null_mut()),
|
||||
next_sibling: AtomicPtr::new(ptr::null_mut()),
|
||||
prev_sibling: AtomicPtr::new(ptr::null_mut()),
|
||||
prev_sibling_or_free_count: PrevSiblingOrFreeCount::new(),
|
||||
next_free: AtomicPtr::new(ptr::null_mut()),
|
||||
free_count: AtomicUsize::new(0),
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -562,9 +578,8 @@ impl RuleNode {
|
|||
refcount: AtomicUsize::new(1),
|
||||
first_child: AtomicPtr::new(ptr::null_mut()),
|
||||
next_sibling: AtomicPtr::new(ptr::null_mut()),
|
||||
prev_sibling: AtomicPtr::new(ptr::null_mut()),
|
||||
prev_sibling_or_free_count: PrevSiblingOrFreeCount::new(),
|
||||
next_free: AtomicPtr::new(FREE_LIST_SENTINEL),
|
||||
free_count: AtomicUsize::new(0),
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -583,6 +598,16 @@ impl RuleNode {
|
|||
}
|
||||
}
|
||||
|
||||
fn prev_sibling(&self) -> &AtomicPtr<RuleNode> {
|
||||
debug_assert!(!self.is_root());
|
||||
unsafe { self.prev_sibling_or_free_count.as_prev_sibling() }
|
||||
}
|
||||
|
||||
fn free_count(&self) -> &AtomicUsize {
|
||||
debug_assert!(self.is_root());
|
||||
unsafe { self.prev_sibling_or_free_count.as_free_count() }
|
||||
}
|
||||
|
||||
/// Remove this rule node from the child list.
|
||||
///
|
||||
/// This method doesn't use proper synchronization, and it's expected to be
|
||||
|
@ -596,7 +621,7 @@ impl RuleNode {
|
|||
// NB: The other siblings we use in this function can also be dead, so
|
||||
// we can't use `get` here, since it asserts.
|
||||
let prev_sibling =
|
||||
self.prev_sibling.swap(ptr::null_mut(), Ordering::Relaxed);
|
||||
self.prev_sibling().swap(ptr::null_mut(), Ordering::Relaxed);
|
||||
|
||||
let next_sibling =
|
||||
self.next_sibling.swap(ptr::null_mut(), Ordering::Relaxed);
|
||||
|
@ -615,7 +640,7 @@ impl RuleNode {
|
|||
// otherwise we're done.
|
||||
if !next_sibling.is_null() {
|
||||
let next = &*next_sibling;
|
||||
next.prev_sibling.store(prev_sibling, Ordering::Relaxed);
|
||||
next.prev_sibling().store(prev_sibling, Ordering::Relaxed);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -760,7 +785,7 @@ impl StrongRuleNode {
|
|||
// only be accessed again in a single-threaded manner when
|
||||
// we're sweeping possibly dead nodes.
|
||||
if let Some(ref l) = last {
|
||||
node.prev_sibling.store(l.ptr(), Ordering::Relaxed);
|
||||
node.prev_sibling().store(l.ptr(), Ordering::Relaxed);
|
||||
}
|
||||
|
||||
return StrongRuleNode::new(node);
|
||||
|
@ -908,14 +933,14 @@ impl StrongRuleNode {
|
|||
let _ = Box::from_raw(weak.ptr());
|
||||
}
|
||||
|
||||
me.free_count.store(0, Ordering::Relaxed);
|
||||
me.free_count().store(0, Ordering::Relaxed);
|
||||
|
||||
debug_assert!(me.next_free.load(Ordering::Relaxed) == FREE_LIST_SENTINEL);
|
||||
}
|
||||
|
||||
unsafe fn maybe_gc(&self) {
|
||||
debug_assert!(self.get().is_root(), "Can't call GC on a non-root node!");
|
||||
if self.get().free_count.load(Ordering::Relaxed) > RULE_TREE_GC_INTERVAL {
|
||||
if self.get().free_count().load(Ordering::Relaxed) > RULE_TREE_GC_INTERVAL {
|
||||
self.gc();
|
||||
}
|
||||
}
|
||||
|
@ -1341,8 +1366,8 @@ impl Drop for StrongRuleNode {
|
|||
|
||||
// Increment the free count. This doesn't need to be an RMU atomic
|
||||
// operation, because the free list is "locked".
|
||||
let old_free_count = root.free_count.load(Ordering::Relaxed);
|
||||
root.free_count.store(old_free_count + 1, Ordering::Relaxed);
|
||||
let old_free_count = root.free_count().load(Ordering::Relaxed);
|
||||
root.free_count().store(old_free_count + 1, Ordering::Relaxed);
|
||||
|
||||
// This can be release because of the locking of the free list, that
|
||||
// ensures that all the other nodes racing with this one are using
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue