Stuff HashMap journal into a string and verify invariants in a few more places.

The RegisterAppMemory approach isn't working for some reason.

MozReview-Commit-ID: GjGaq6GALI5
This commit is contained in:
Bobby Holley 2017-10-09 15:52:49 -07:00
parent b1926b4656
commit 4b64390b81
3 changed files with 65 additions and 27 deletions

View file

@ -1,7 +1,6 @@
use hash_map::HashMap; use hash_map::HashMap;
use std::borrow::Borrow; use std::borrow::Borrow;
use std::hash::{BuildHasher, Hash}; use std::hash::{BuildHasher, Hash};
use table::SafeHash;
use FailedAllocationError; use FailedAllocationError;
@ -12,9 +11,9 @@ const CANARY: usize = 0x42cafe9942cafe99;
#[derive(Clone, Debug)] #[derive(Clone, Debug)]
enum JournalEntry { enum JournalEntry {
Insert(SafeHash), Insert(usize),
GetOrInsertWith(SafeHash), GOIW(usize),
Remove(SafeHash), Remove(usize),
DidClear(usize), DidClear(usize),
} }
@ -37,17 +36,23 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
&self.map &self.map
} }
#[inline(always)] #[inline(never)]
pub fn begin_mutation(&mut self) { pub fn begin_mutation(&mut self) {
self.map.verify();
assert!(self.readonly); assert!(self.readonly);
self.readonly = false; self.readonly = false;
self.verify();
} }
#[inline(always)] #[inline(never)]
pub fn end_mutation(&mut self) { pub fn end_mutation(&mut self) {
self.map.verify();
assert!(!self.readonly); assert!(!self.readonly);
self.readonly = true; self.readonly = true;
self.verify();
}
fn verify(&self) {
let mut position = 0; let mut position = 0;
let mut bad_canary: Option<(usize, *const usize)> = None; let mut bad_canary: Option<(usize, *const usize)> = None;
for (_,v) in self.map.iter() { for (_,v) in self.map.iter() {
@ -105,7 +110,7 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
default: F default: F
) -> Result<&mut V, FailedAllocationError> { ) -> Result<&mut V, FailedAllocationError> {
assert!(!self.readonly); assert!(!self.readonly);
self.journal.push(JournalEntry::GetOrInsertWith(self.map.make_hash(&key))); self.journal.push(JournalEntry::GOIW(self.map.make_hash(&key).inspect()));
let entry = self.map.try_entry(key)?; let entry = self.map.try_entry(key)?;
Ok(&mut entry.or_insert_with(|| (CANARY, default())).1) Ok(&mut entry.or_insert_with(|| (CANARY, default())).1)
} }
@ -113,7 +118,7 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
#[inline(always)] #[inline(always)]
pub fn try_insert(&mut self, k: K, v: V) -> Result<Option<V>, FailedAllocationError> { pub fn try_insert(&mut self, k: K, v: V) -> Result<Option<V>, FailedAllocationError> {
assert!(!self.readonly); assert!(!self.readonly);
self.journal.push(JournalEntry::Insert(self.map.make_hash(&k))); self.journal.push(JournalEntry::Insert(self.map.make_hash(&k).inspect()));
let old = self.map.try_insert(k, (CANARY, v))?; let old = self.map.try_insert(k, (CANARY, v))?;
Ok(old.map(|x| x.1)) Ok(old.map(|x| x.1))
} }
@ -124,7 +129,7 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
Q: Hash + Eq Q: Hash + Eq
{ {
assert!(!self.readonly); assert!(!self.readonly);
self.journal.push(JournalEntry::Remove(self.map.make_hash(k))); self.journal.push(JournalEntry::Remove(self.map.make_hash(k).inspect()));
self.map.remove(k).map(|x| x.1) self.map.remove(k).map(|x| x.1)
} }
@ -141,19 +146,22 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
#[inline(never)] #[inline(never)]
fn report_corruption( fn report_corruption(
&mut self, &self,
canary: usize, canary: usize,
canary_addr: *const usize, canary_addr: *const usize,
position: usize position: usize
) { ) {
use ::std::ffi::CString;
let key = b"HashMapJournal\0";
let value = CString::new(format!("{:?}", self.journal)).unwrap();
unsafe { unsafe {
Gecko_AddBufferToCrashReport( Gecko_AnnotateCrashReport(
self.journal.as_ptr() as *const _, key.as_ptr() as *const ::std::os::raw::c_char,
self.journal.len() * ::std::mem::size_of::<JournalEntry>(), value.as_ptr(),
); );
} }
panic!( panic!(
"HashMap Corruption (sz={}, cap={}, pairsz={}, cnry={:#x}, pos={}, base_addr={:?}, cnry_addr={:?})", "HashMap Corruption (sz={}, cap={}, pairsz={}, cnry={:#x}, pos={}, base_addr={:?}, cnry_addr={:?}, jrnl_len={})",
self.map.len(), self.map.len(),
self.map.raw_capacity(), self.map.raw_capacity(),
::std::mem::size_of::<(K, (usize, V))>(), ::std::mem::size_of::<(K, (usize, V))>(),
@ -161,6 +169,7 @@ impl<K: Hash + Eq, V, S: BuildHasher> DiagnosticHashMap<K, V, S>
position, position,
self.map.raw_buffer(), self.map.raw_buffer(),
canary_addr, canary_addr,
self.journal.len(),
); );
} }
} }
@ -205,6 +214,6 @@ impl<K: Hash + Eq, V, S: BuildHasher> Drop for DiagnosticHashMap<K, V, S>
} }
extern "C" { extern "C" {
pub fn Gecko_AddBufferToCrashReport(addr: *const ::std::os::raw::c_void, pub fn Gecko_AnnotateCrashReport(key_str: *const ::std::os::raw::c_char,
bytes: usize); value_str: *const ::std::os::raw::c_char);
} }

View file

@ -694,6 +694,12 @@ impl<K, V, S> HashMap<K, V, S>
self.table.raw_buffer() self.table.raw_buffer()
} }
/// Verify that the table metadata is internally consistent.
#[inline]
pub fn verify(&self) {
self.table.verify();
}
/// Reserves capacity for at least `additional` more elements to be inserted /// Reserves capacity for at least `additional` more elements to be inserted
/// in the `HashMap`. The collection may reserve more space to avoid /// in the `HashMap`. The collection may reserve more space to avoid
/// frequent reallocations. /// frequent reallocations.

View file

@ -247,13 +247,21 @@ impl<K, V> RawBucket<K, V> {
(self.hash(), self.pair()) (self.hash(), self.pair())
} }
fn assert_bounds(&self, bytes_allocated: usize) { fn assert_bounds(&self, bytes_allocated: usize, size: Option<usize>) {
let base = self.hash_start as *mut u8; let base = self.hash_start as *mut u8;
let (h, p) = unsafe { self.hash_pair() }; let (h, p) = unsafe { self.hash_pair() };
assert!((h as *mut u8) < (p as *mut u8), "HashMap Corruption - hash offset not below pair offset"); assert!((h as *mut u8) < (p as *mut u8), "HashMap Corruption - hash offset not below pair offset");
let end = unsafe { p.offset(1) } as *mut u8; let end = unsafe { p.offset(1) } as *mut u8;
assert!(end > base, "HashMap Corruption - end={:?}, base={:?}", end, base); assert!(end > base, "HashMap Corruption - end={:?}, base={:?}, idx={}, alloc={}, size={:?}", end, base, self.idx, bytes_allocated, size);
assert!(end <= unsafe { base.offset(bytes_allocated as isize) }, "HashMap Corruption - end={:?}, base={:?}", end, base); assert!(
end <= unsafe { base.offset(bytes_allocated as isize) },
"HashMap Corruption - end={:?}, base={:?}, idx={}, alloc={}, size={:?}",
end,
base,
self.idx,
bytes_allocated,
size,
);
} }
} }
@ -431,13 +439,13 @@ impl<K, V, M: Deref<Target = RawTable<K, V>>> Bucket<K, V, M> {
/// Modifies the bucket in place to make it point to the next slot. /// Modifies the bucket in place to make it point to the next slot.
pub fn next(&mut self) { pub fn next(&mut self) {
self.raw.idx = self.raw.idx.wrapping_add(1) & self.table.capacity_mask; self.raw.idx = self.raw.idx.wrapping_add(1) & self.table.capacity_mask;
self.raw.assert_bounds(self.table.bytes_allocated); self.raw.assert_bounds(self.table.bytes_allocated, None);
} }
/// Modifies the bucket in place to make it point to the previous slot. /// Modifies the bucket in place to make it point to the previous slot.
pub fn prev(&mut self) { pub fn prev(&mut self) {
self.raw.idx = self.raw.idx.wrapping_sub(1) & self.table.capacity_mask; self.raw.idx = self.raw.idx.wrapping_sub(1) & self.table.capacity_mask;
self.raw.assert_bounds(self.table.bytes_allocated); self.raw.assert_bounds(self.table.bytes_allocated, None);
} }
} }
@ -813,6 +821,7 @@ impl<K, V> RawTable<K, V> {
} }
fn raw_bucket_at(&self, index: usize) -> RawBucket<K, V> { fn raw_bucket_at(&self, index: usize) -> RawBucket<K, V> {
self.verify();
let hashes_size = self.capacity() * size_of::<HashUint>(); let hashes_size = self.capacity() * size_of::<HashUint>();
let pairs_size = self.capacity() * size_of::<(K, V)>(); let pairs_size = self.capacity() * size_of::<(K, V)>();
@ -833,7 +842,7 @@ impl<K, V> RawTable<K, V> {
} }
}; };
bucket.assert_bounds(self.bytes_allocated); bucket.assert_bounds(self.bytes_allocated, Some(self.size));
bucket bucket
} }
@ -843,6 +852,20 @@ impl<K, V> RawTable<K, V> {
self.hashes.ptr() as *const u8 self.hashes.ptr() as *const u8
} }
/// Verify that the table metadata is internally consistent.
#[inline]
pub fn verify(&self) {
assert!(
self.capacity() == 0 || self.capacity().is_power_of_two(),
"HashMap Corruption: mask={}, sz={}, alloc={}", self.capacity_mask, self.size, self.bytes_allocated,
);
assert_eq!(
self.capacity() * (size_of::<usize>() + size_of::<(K, V)>()),
self.bytes_allocated,
"HashMap Corruption: mask={}, sz={}, alloc={}", self.capacity_mask, self.size, self.bytes_allocated,
);
}
/// Creates a new raw table from a given capacity. All buckets are /// Creates a new raw table from a given capacity. All buckets are
/// initially empty. /// initially empty.
pub fn new(capacity: usize) -> Result<RawTable<K, V>, FailedAllocationError> { pub fn new(capacity: usize) -> Result<RawTable<K, V>, FailedAllocationError> {
@ -933,7 +956,7 @@ impl<K, V> RawTable<K, V> {
} }
} }
raw.idx = raw.idx.checked_sub(1).unwrap(); raw.idx = raw.idx.checked_sub(1).unwrap();
raw.assert_bounds(self.bytes_allocated); raw.assert_bounds(self.bytes_allocated, Some(self.size));
} }
} }
@ -993,12 +1016,12 @@ impl<'a, K, V> Iterator for RawBuckets<'a, K, V> {
self.elems_left = self.elems_left.checked_sub(1).unwrap(); self.elems_left = self.elems_left.checked_sub(1).unwrap();
if self.elems_left != 0 { if self.elems_left != 0 {
self.raw.as_mut().unwrap().idx += 1; self.raw.as_mut().unwrap().idx += 1;
self.raw.as_ref().unwrap().assert_bounds(self.bytes_allocated); self.raw.as_ref().unwrap().assert_bounds(self.bytes_allocated, None);
} }
return Some(item); return Some(item);
} }
self.raw.as_mut().unwrap().idx += 1; self.raw.as_mut().unwrap().idx += 1;
self.raw.as_ref().unwrap().assert_bounds(self.bytes_allocated); self.raw.as_ref().unwrap().assert_bounds(self.bytes_allocated, None);
} }
} }
} }
@ -1207,9 +1230,9 @@ impl<K: Clone, V: Clone> Clone for RawTable<K, V> {
} }
buckets.idx += 1; buckets.idx += 1;
buckets.assert_bounds(self.bytes_allocated); buckets.assert_bounds(self.bytes_allocated, None);
new_buckets.idx += 1; new_buckets.idx += 1;
new_buckets.assert_bounds(new_ht.bytes_allocated); new_buckets.assert_bounds(new_ht.bytes_allocated, None);
} }
new_ht.size = self.size(); new_ht.size = self.size();